Skip to content

sukeshan/CineScript-Insight-Agent

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

15 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Script Analysis System

An agentic AI system designed to analyze short-form scripts and generate structured storytelling insights through a unified, memory-aware pipeline.

What the Agent Can Do

This system continuously processes and reasons about script content using an intelligent loop:

  1. Upload & Initial Insight: Upon script upload, a summary is generated instantly. Simultaneously, deep analysis for characters, entity mapping, and scene splitting run in the background.
  2. Stateful Conversation: A ReAct (Reasoning and Acting) agent powered by LangGraph maintains conversation memory. It autonomously determines when to use specialized tools to pull deep insights from pre-generated analysis files to answer your questions.
  3. Context Budget Management: Automated context engineering manages the LLM context window. When the budget hits 40% of the model's limit, the system compresses the conversation history and drops the raw script to maintain optimal performance and reduce costs.

Core Design Principles

  • Progressive Disclosure: The system is built to provide immediate value while doing heavy lifting in the background. The script summary is shown instantly to the user, while background analysis completes silently. Deeper insights are retrieved on-demand as the user asks specific questions.
  • Shared KV-Cache Prefix: Background tasks (Summary, Character Analysis, Entity Mapping, Scene Splitting) share a common system prompt and script prefix to maximize AI prompt caching performance, significantly reducing latency and cost.
  • Parallel Processing: Initial background tasks execute concurrently using asyncio.
  • Skill-based Architecture & Unified Context: Generated analysis files (e.g., character traits, scene entities) are treated as modular "skills." The agent dynamically loads only the required skill files into its working memory when needed to answer a user's question, rather than stuffing the entire context at once.
  • Interactive Feedback: Every agent response generates relevant follow-up "chips" to guide the user deeper into the analysis.

System Flow

flowchart TD
    Upload(["🎬 Script Upload\nStreamlit file input"]):::gray --> Pre["Validate & Preprocess\nextract text · metadata"]:::gray
    Pre --> Lock(["🔒 UI Locked\ntyping bar hidden"]):::gray
    Lock --> Prefix["Shared Prefix Builder\nsystem: static analyst prompt\nuser: script text\nassistant: Script received\nKV Cache Boundary"]:::teal

    Prefix -->|"seed cache"| SA["Summary Agent"]:::purple
    Prefix -->|"Background Parallel"| CA["Character Analyst"]:::coral
    Prefix -->|"Background Parallel"| EM["Entity Mapper"]:::amber
    Prefix -->|"Background Parallel"| SS["Scene Splitter"]:::teal

    SA -->|"resolves first"| SUM(["✅ Summary shown\n+ follow-up questions\ntyping bar unlocked"]):::purple

    CA -->|"wait"| SI["Skills Index Builder"]:::teal
    EM -->|"wait"| SI
    SS -->|"wait"| SI

    SI --> Fin["Finalize Context"]:::teal
    Fin --> SC[("ScriptContext\n.md files ready")]:::teal

    SUM -->|"user asks question"| Agent["LangGraph ReAct Agent\nUnified State Memory"]:::blue
    SC -->|"file paths registered"| Agent

    Agent -->|"needs data"| Tool{"execute_tool"}:::amber
    Tool -->|"load_file / write_file"| Agent
    Agent -->|"final_answer"| Response(["💬 Response\n+ follow-up questions"]):::gray

    Response --> Budget{"Context at\n40% of limit?"}:::amber
    Budget -->|"no"| Agent
    Budget -->|"yes"| Compress["Context Compression\nRemove raw script from prompt\nCompress older turns"]:::coral
    Compress --> Agent

    classDef gray   fill:#888780,stroke:#5F5E5A,color:#fff
    classDef teal   fill:#1D9E75,stroke:#0F6E56,color:#fff
    classDef purple fill:#7F77DD,stroke:#534AB7,color:#fff
    classDef coral  fill:#D85A30,stroke:#993C1D,color:#fff
    classDef amber  fill:#BA7517,stroke:#854F0B,color:#fff
    classDef blue   fill:#185FA5,stroke:#0C447C,color:#fff
Loading

Project Structure

script-analysis/
│
├── agents/                  # LangGraph nodes
│   ├── summary.py           # Initial Summary Generator
│   ├── characters.py        # Character Analyst
│   ├── entity_mapper.py     # Entity Mapper
│   ├── scene_splitter.py    # Script Scene Fragmenter
│   ├── skills_index_builder.py # Skills/Files Indexer
│   └── conversation_graph.py# Stateful ReAct Agent (LangGraph)
│
├── core/                    # Core logic & utilities
│   ├── context.py           # ScriptContext source of truth
│   ├── pipeline.py          # Pipeline orchestration
│   ├── context_manager.py   # 40% threshold compression logic
│   ├── llm.py               # Structured output & token tracking
│   ├── prompts.py           # Dynamic system prompt builder
│   └── tools.py             # Agent tools (load_file, write_file)
│
├── models/                  # Pydantic output schemas
│   ├── entities.py
│   ├── characters.py
│   └── summary.py
│
├── ui/                      # Frontend
│   └── app.py               # Streamlit Dashboard & Chat Interface
│
├── outputs/                 # Analysis assets (auto-generated)
│   ├── character_analysis.md
│   ├── entity_map.md
│   └── skills_index.md
│
└── requirements.txt

Local Setup Guidelines

Follow these steps to run the Script Analysis System locally:

  1. Clone the repository

    git clone <repository-url>
    cd content_agent
  2. Environment Setup & Dependencies (using uv) We recommend using uv for lightning-fast environment setup:

    # Install uv (if you haven't already)
    curl -LsSf https://astral.sh/uv/install.sh | sh
    
    # Create a virtual environment
    uv venv
    
    # Activate the environment
    source .venv/bin/activate  # On Windows use: .venv\Scripts\activate
    
    # Install dependencies
    uv pip install -r requirements.txt

    (Alternatively, you can use standard python -m venv venv and pip install -r requirements.txt)

  3. Environment Variables Create a .env file in the root directory and add your OpenAI API key:

    OPENAI_API_KEY=your-api-key-here
  4. Run the Application Launch the Streamlit UI:

    streamlit run ui/app.py

Future Directions & Improvements

To further enhance the capabilities and robust nature of the agent, the following architectural and feature updates are planned:

  • MCP Tool Server integration: Transition the tool execution layer to act as a Model Context Protocol (MCP) server, standardizing how context and tools are provided.
  • Long-term Memory: Implement persistent cross-session memory (e.g., SQLite or Redis) allowing the agent to remember user preferences, previous script discussions, and ongoing narrative arcs.
  • Prompt Density & Preference Optimization: Refine prompts to increase information density and implement preference optimization (like DPO/RLHF algorithms) to better align with user styles.
  • Proper Observation and Log Traces: Integrate tracing tools (such as LangSmith, Phoenix, or OpenTelemetry) to monitor the LangGraph ReAct loops, token usage, tool failure rates, and agent reasoning paths for debugging and analytics.
  • Chunking + Map-Reduce: Scalable ingestion for feature-length scripts.
  • Streaming Tokens: Stream tokens directly from each Phase 2 agent to Streamlit surfaces for lower perceived latency.
  • Multi-Script Comparison Mode: Compare two drafts side-by-side to show what changed in the character arcs or engagement.
  • Fine-tuned Classifier for Entities: Train a small, lightweight classifier for entity type labeling to completely remove the risk of LLM hallucinations in structured data.
  • Confidence Scores: Calculate and display confidence scores on engagement factors and emotion labels.

About

An agentic AI system designed to analyze short-form scripts and generate structured storytelling insights through a unified, memory-aware pipeline

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages