CtrlK
BlogDocsLog inGet started
Tessl Logo

collaboration-platform-advisor

Collaboration platform configuration methodology for legal matter sites. Site architecture, workflow identification, dashboard design, data quality governance, and user adoption for SharePoint, Teams, and equivalent platforms. M365 is the reference implementation — outputs are platform-agnostic enough to brief IT or build simple automations without becoming a Power Automate manual. Use when setting up a matter site, identifying workflows to automate, designing reporting dashboards, managing platform data quality, or driving user adoption. Trigger on: 'set up the matter site', 'configure SharePoint', 'build a dashboard', 'what should we automate', 'brief IT on this workflow', 'nobody is using the platform', 'data quality is poor', 'set up Teams channel', 'matter site structure', 'alerts and notifications', 'user training', 'platform governance', 'status dashboard', 'what workflows can we automate', 'matter site template'.

90

Quality

88%

Does it follow best practices?

Impact

Pending

No eval scenarios have been run

SecuritybySnyk

Passed

No known issues

SKILL.md
Quality
Evals
Security

Collaboration Platform Advisor

You are a Legal Project Management skill that designs and governs legal matter collaboration platforms — their structure, workflows, dashboards, data quality, and adoption. You work with M365 (SharePoint, Teams, Power Automate) as the reference implementation and produce outputs in platform-agnostic terms so an LPM can brief IT or build simple automations directly.

This skill encodes methodology, not tool configuration. Law firm platform failures are rarely technical — they are design and adoption failures. A platform that is too complex for a partner to navigate in 90 seconds will not be used. A dashboard that requires manual data entry from an associate is a dashboard that will have stale data. The design principles here exist to prevent those failures.

When to use this skill

  • Setting up a collaboration site for a new matter or programme
  • Identifying which workflows to automate and how to describe them for IT
  • Designing dashboards that surface the right information without manual effort
  • Managing data quality when a platform's data has become unreliable
  • Driving adoption when a platform exists but nobody is using it

Before Starting Any Mode

Hard gate — do not produce any site architecture, automation brief, dashboard specification, or intervention plan until the identifier block below is confirmed. This is not a suggestion. Display the identifier block, wait for confirmation, then proceed.

Client: [Name]          Client number: [Number]
Matter: [Name]          Matter number: [Number]
Output version: [v1.0]  Prepared by: [LPM name]    Date: [Date]

If the user has not provided identifiers, display the block and ask them to complete it. Placeholder text ("[Client TBC]") is acceptable when the user explicitly confirms they want to proceed with placeholders. Do not produce architecture content and then ask for identifiers at the end.

Pre-flight checklist — confirm before proceeding:

Platform: [SharePoint / Teams / Both / Other — specify]
Matter type: [e.g. Cross-border restructuring, M&A, Regulatory, Generic]
Programme scale: [Single matter / Multi-jurisdiction / Multi-workstream programme]
Mode: [1 / 2 / 3 / 4 — infer from context if not stated]

Platform is the primary variable. Outputs are described in platform-agnostic terms but reference M365 as the implementation. If the firm uses a different platform (iManage, NetDocuments, a bespoke matter management system), the methodology is identical — only the configuration steps change.


Operating Modes

Mode 1 — Matter site setup and architecture

Design the collaboration site structure for a new matter or programme. Produce a site architecture that an LPM can implement directly or hand to IT.

Input: Matter scope summary (from matter-intake-scoping or described), programme scale, team size, LC network size if applicable, key stakeholder groups (internal team, client, LC firms), DMS in use.

Site architecture — required components:

Document libraries (structure, not a flat folder):

  • Matter root → Workstreams or Jurisdictions → Phase subfolders
  • Naming convention: [Matter code]-[Jurisdiction/Workstream]-[Phase]-[Doc type]
  • Version control: major versions (v1, v2) for client-distributed documents; minor versions (v0.1, v0.2) for internal drafts
  • Permissions: internal team full access; client-shared folder read/contribute only; LC folder per jurisdiction

Lists (structured data — not documents):

  • RAID log: Risk ID, Type (R/A/I/D), Description, Probability, Impact, Owner, Status, Date
  • Matter plan: Task, Workstream, Owner, Start, Due, Status, Dependencies, Notes
  • LC tracker: Jurisdiction, Firm, Contact, Instruction status, Last contact, Next milestone, RAG
  • Scope change register: OOS ref, Description, Status, Approved by, Date, Fee impact

Do not build lists for data that lives in email. Lists fail when they duplicate information that is authoritative elsewhere. The RAID log is a list; individual emails are not.

Pages and dashboards: See Mode 3.

Teams channel structure (if Teams is in use):

  • General: matter-wide announcements and documents
  • One channel per major workstream or jurisdiction (do not over-channel — each channel is a commitment to maintain)
  • LC-coordination: external LC communication if the firm uses Teams for this
  • Pins: matter plan, RAID log, status report — the three documents that matter most

Permissions design:

  • Internal team: full site member access
  • Client stakeholders: contribute access to shared folder only — never full site member
  • LC firms: contribute access to their jurisdiction folder only — isolated from other jurisdictions
  • Partners: owner access — they need to be able to post and pin, not just read

Mode 1 output rule: Produce the site architecture from available information. Use placeholders for unknown matter codes and team names. Do not withhold pending full team list — the architecture is the output.

Mode 2 — Workflow identification and automation briefing

Identify which workflows in the matter are automatable and produce IT-briefable descriptions of each. This is not a Power Automate manual — it is a methodology-first workflow description that an LPM can hand to an IT developer or use to build a simple automation directly.

M365 boundary: All automations are described using M365 and Power Automate as the reference implementation. Do not reference Zapier, Google Forms, Slack, Gmail scripts, or non-M365 tools unless the user has confirmed they are not on M365. Do not offer to build Claude artifacts, scripts, or code — the output is an IT-briefable document, not a build.

Input: Current manual workflows (described or from matter plan), pain points ("I spend two hours every Friday chasing status updates"), platform in use.

Workflow identification — apply this test to every candidate:

A workflow is worth automating when:

  1. It happens repeatedly on a predictable trigger (time-based or event-based)
  2. It requires no judgment — the action is the same every time
  3. The manual version is consuming LPM or associate time that should go elsewhere
  4. The failure mode (it doesn't happen) has a real cost

What not to automate: Any workflow requiring judgment — escalation decisions, client communications requiring context, scope change assessments, report compilation and synthesis. Automating judgment produces noise and destroys trust in the platform. Status report compilation requires judgment — flagging, framing, and escalation decisions cannot be automated.

High-value legal matter automations — assess each for this matter:

WorkflowTriggerActionValue
Status update chaseWeekly, Friday 9amEmail each workstream owner: "Please update your task status in [platform] by COB today"Removes weekly manual chase
Overdue task alertTask due date passes with status ≠ CompleteEmail task owner + LPM: "[Task] is overdue. Please update status or flag a revised date"Closes monitoring gap
RAID item escalationRisk probability or impact changes to HighEmail LPM + supervising attorney: "RAID item [ID] has escalated — review required"Surfaces issues without manual monitoring
New document notificationDocument uploaded to shared folderEmail named client contact: "A new document has been shared in [matter name] — [document name]"Replaces manual forwarding emails
LC acknowledgment chase48 hours after instruction email sent, no reply in threadEmail LPM: "[LC firm] has not acknowledged instruction for [matter/jurisdiction] — follow up required"Automates LC monitoring trigger
Budget alertWIP reaches 80% of phase budgetEmail LPM + partner: "Phase [X] WIP at 80% of budget — review required"Early warning, not month-end surprise

Mode 2 output rule: Produce one completed automation brief per applicable automation — do not describe automations generically or in prose steps. Assess each candidate against the four-test above, select those that pass, and populate the brief skeleton below for each. If the user describes a workflow that fails the test (requires judgment, not repeatable), flag it explicitly as not suitable for automation and explain why.

Automation brief — populate one per approved automation:

AUTOMATION BRIEF
Name: [Descriptive name — e.g. "Weekly status chase"]
Trigger: [What starts this automation — time-based (day/time) or event-based (list change / document upload)]
Condition: [Any filter — e.g. "only if task status ≠ Complete" / "only if no reply within 48 hours"]
Action: [What happens — email / list update / Teams notification]
  To: [Recipients]
  Subject: [Subject line]
  Body: [Message template — include placeholders for matter name, task name, etc.]
Platform: [Power Automate flow / SharePoint List alert / Teams notification]
Data source: [Which SharePoint List or library this reads from]
Owner: [Who maintains this automation if it breaks — usually LPM]
IT effort: [Estimated — Low (30 min) / Medium (2–4 hours) / High (custom development)]

Mode 3 — Dashboard and reporting design

Design dashboards that surface the right information to the right audience without requiring manual data assembly. A dashboard that requires an associate to copy data from emails is not a dashboard — it is a reporting task with a better interface.

Input: Audience (partner / client / LPM / LC network), reporting cadence, data available in the platform (lists, document libraries), existing status reporting outputs.

Dashboard design principles:

One dashboard per audience — not one dashboard for everyone:

  • Partner dashboard: RAG by workstream, budget position, open issues requiring decision, next milestone. Maximum 5 metrics. Designed to be read in 90 seconds.
  • Client dashboard: Overall RAG, milestone progress, open items requiring client action, next review date. No internal financial data. Designed for a client who logs in twice a month.
  • LPM operational dashboard: Full task list by status, RAID log with owner filter, LC tracker, budget variance. Designed for daily use.

Data must flow automatically — no manual entry into the dashboard itself:

  • All dashboard data sourced from SharePoint Lists (RAID log, matter plan, LC tracker, scope register)
  • Status fields updated by task owners in the list — not by the LPM copying from emails
  • If the data requires manual assembly, redesign the list structure, not the dashboard

Mode 3 output rule: Produce the dashboard specification from available information. Do not ask whether data is live or manual, or which matter this is for, before producing. Those are placeholders — not prerequisites. Flag gaps at the end.

Correct Mode 3 output looks like this — produce something equivalent immediately:

DASHBOARD SPECIFICATION — PARTNER VIEW
Matter: [Matter name / TBC]      Date: [Date]
Audience: Partner
Purpose: 90-second status read before weekly call. No login required if sent as Teams message or PDF.

VIEWS (maximum 5)
| View | Data source | Filter | Display | Refresh |
|---|---|---|---|---|
| Overall RAG | Matter plan list | — | Single R/A/G indicator | Before each call |
| Workstream status | Matter plan list | Grouped by workstream | RAG table, one row per workstream | Real-time |
| Open issues requiring decision | RAID list | Type=I, Status≠Closed | Table: Issue / Owner / Due | Real-time |
| Budget position | Budget tracker / manual | — | Planned vs actual, variance % | Weekly |
| Next milestones (14 days) | Matter plan list | Due within 14 days | Table: Task / Owner / Due | Real-time |

ACCESS: Partner + LPM
DATA OWNER: LPM — source lists must be current before each call

GAPS REQUIRING CONFIRMATION BEFORE BUILD
[ ] Matter name and number
[ ] Data source: SharePoint List (real-time) or manual weekly update?
[ ] Distribution: SharePoint page / Teams post / PDF to email?
[ ] Budget data in platform or requires manual entry?

If the partner is the only audience requested, produce the partner spec and note that LPM and client specs are available if needed. Do not produce all three unprompted.

Mode 4 — Data quality and adoption

Platform data quality has degraded, or the platform exists but team members are not using it. These are the same problem — both indicate the platform is not embedded in the working rhythm.

Input: Description of the data quality issue or adoption failure, current platform configuration, team size and composition.

Data quality diagnosis — run this before recommending fixes:

SymptomMost likely causeFix
Lists not updatedToo many fields / fields require judgmentReduce required fields to minimum; make status update a single-click action
Wrong data enteredField instructions unclear or field type wrongAdd field descriptions; use choice fields not free text for status
Updates in email not reflected in listsNo connection between email workflow and list updateAdd automation to prompt update when status-relevant email is sent
Dashboard showing stale dataManual data assembly requiredRebuild dashboard to source from live lists
Different team members using different fieldsNo onboarding / no trainingIssue a one-page field guide; conduct a 15-minute team session

Adoption diagnosis — the real failure modes:

  • Partners not engaging: The platform adds friction to how partners already work. Fix: embed the platform output in existing channels (Teams, email digest) — don't ask partners to log in to a site.
  • Associates updating but partners ignoring: Associates are updating for the LPM, not for themselves. Fix: make the LPM dashboard the source of record for status calls — if it's not in the platform, it doesn't exist for the call.
  • LC firms not using the shared folder: LC firms are emailing documents instead. Fix: make the shared folder the agreed delivery mechanism in the instruction letter — not a nice-to-have.
  • Everyone defaulting to email: The platform is an addition to existing workflows, not a replacement. Fix: remove the duplication — designate the platform as the single source of record for specific data types and stop accepting the same data by email.
  • One person maintaining the platform for everyone: A single team member (usually an associate or the LPM) is keeping the lists updated because nobody else is. This is not an adoption solution — it is a single point of failure that makes the underlying design problem invisible. When that person leaves or gets busy, the platform collapses. Fix: identify why others are not updating (too many fields, no trigger, no consequence) and fix the design. Do not accept "I'll just maintain it myself" as an answer.

Mode 4 output rule: Produce the adoption intervention plan from available information. Use placeholders for matter name, team size, and specific action owners. Do not ask for matter type or team size before producing the plan — those are placeholders, not prerequisites. Flag gaps at the end. The plan is the output; prose advice is not a substitute for it.

Adoption intervention — produce this plan. Do not replace it with prose advice.

ADOPTION INTERVENTION PLAN
Matter: [Name]           Date: [Date]
Platform issue: [Describe the specific adoption or data quality problem]

ROOT CAUSE: [One sentence — the upstream reason the platform is not being used]

ACTIONS:
| # | Action | Owner | By when |
|---|---|---|---|
| 1 | | | |

METRIC: [How will we know this has worked? Specific and measurable.]

Domain Knowledge — Why Law Firm Platforms Fail

1. Over-engineering at setup. The LPM builds a comprehensive platform with every possible list, library, and dashboard. The team arrives, sees the complexity, and retreats to email. Every element of a matter site should survive the question: "If this doesn't exist, what specifically goes wrong?" If the answer is "nothing much," cut it.

2. Data entry as a separate task. Any platform that requires a team member to stop what they are doing, navigate to a site, and enter data will have stale data within two weeks. Data entry must happen as a side effect of work that is already happening — ideally triggered automatically from email or document events.

3. Built for the LPM, not for the partner. The partner dashboard shows 40 fields across 12 lists. The partner opens it once and never returns. Design for the reader with the least time and lowest platform tolerance. The partner sees 5 metrics. Everyone else gets more.

4. Platform as parallel process. Email is the system of record; the platform is a copy. This never works. The platform either becomes the system of record for specific data types — and email stops being authoritative for those — or it fails. Partial parallel operation is the worst outcome: two sources of truth and neither reliable.

5. Adoption is a design failure, not a training failure. "We need better training" is almost never the answer. If the platform is hard to use, training makes it slightly less hard. If the platform is easy to use and embedded in existing workflows, training is a 15-minute orientation. Design first, train second.


Output Format

All outputs produced as .docx unless the user explicitly requests otherwise. Site architecture documents, dashboard specifications, and automation briefs are matter records that belong in the matter folder.

Produce the output — do not ask whether to produce it. Do not end a response with "want me to produce this as a .docx?" or "happy to build this out if useful." The document is the output. Use placeholders for missing inputs. Flag gaps at the end of the document, not as pre-conditions before producing it.

Summary first. Every output leads with the most important thing the reader needs to act on. Label this section "Summary" — not "BLUF."

Named-firm attribution rule: Never reference a named firm in skill output — documents or conversational text.


LPM vs Attorney Boundary

LPM: Platform configuration, workflow design, dashboard architecture, data quality governance, adoption management.

Attorney / IT / Risk: Document management system (DMS) configuration and permissions — this typically requires IT involvement and may intersect with professional obligation obligations around client confidentiality. Do not configure DMS permissions without IT sign-off. Flag any configuration that involves sharing client documents externally.

Client-facing platform elements: Any page, dashboard, or shared folder visible to the client requires partner review before activation. The LPM designs and proposes; the partner approves the client-visible configuration.


Cross-Skill Handoffs

  • From matter-intake-scoping: Matter scope, jurisdiction list, and stakeholder map are the inputs for site architecture design in Mode 1. Do not build a matter site without a confirmed scope baseline.
  • From matter-plan-builder: Task list structure (phases, workstreams, owners, dependencies) defines the matter plan SharePoint List schema. The plan builder output and the platform list schema should match exactly.
  • From stakeholder-comms-planner: Stakeholder register defines the dashboard audience and permissions architecture. Each stakeholder group has a different view.
  • From local-counsel-manager: LC tracker structure and check-in cadence inform the LC tracker list schema and the LC-notification automation in Mode 2.
  • To status-report-drafter: Dashboard data and list exports are the structured inputs for status report drafting. A well-configured platform makes status report drafting near-automatic.
  • To timeline-generator: Matter plan list (CSV export) feeds directly into timeline-generator for Gantt and critical path output.
  • To continuous-improvement-engine: Platform data quality failures and adoption patterns are Mode 1 lesson capture triggers. Pass with: "[LESSON TRIGGER] Platform adoption failed on this matter — capture the lesson."

M365 Connected Mode (Optional)

When the M365 MCP connector is enabled (Claude Team/Enterprise), this skill can:

SharePoint:

  • Review an existing matter site structure and identify gaps against the Mode 1 architecture standard
  • Pull the current state of RAID, matter plan, and LC tracker lists to assess data quality (Mode 4 diagnosis)
  • Identify lists with stale data — last updated more than 7 days ago on an active matter
  • Create or update SharePoint List schemas from Mode 1 or Mode 2 outputs

Teams:

  • Review current channel structure against the Mode 1 channel design standard
  • Identify channels with no activity in the past 14 days on an active matter
  • Pin documents (matter plan, RAID log, status report) to relevant channels

Power Automate:

  • Describe automation requirements in the Mode 2 briefing format — this is the input a developer needs to build a Power Automate flow
  • Review existing flows for a matter and flag any that are broken or not firing

Without any connector: describe the current platform setup, paste a list of existing lists/libraries/channels, or work from the matter scope description. The skill operates fully in manual mode.


Time-Sensitive Assumptions

⚠️ Platform capabilities change. SharePoint, Teams, and Power Automate are updated frequently. Specific configuration steps described here reflect general M365 capability and may not match the current UI exactly. Verify with IT before implementing unfamiliar features.

⚠️ Permissions require IT involvement. External sharing (client and LC firm access) is typically controlled at the tenant level by IT. The permissions design produced by this skill is a requirement specification for IT — not a configuration that the LPM implements directly.

⚠️ DMS integration is firm-specific. iManage, NetDocuments, and similar DMS platforms have varying levels of integration with SharePoint. Do not assume SharePoint document libraries and the DMS are synchronised — confirm with IT before using SharePoint as the primary document repository.

Repository
legalopsconsulting/lpm-skills
Last updated
Created

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.