LabFlow AI pipelines 6 repetitive research tasks through an agentic LLM system — extracting structured insights, comparing studies, and generating reports without manual effort.
Each workflow is an OpenAI tool. The agent selects and chains them automatically based on your task — no manual routing.
Transforms raw research notes into structured summaries — objectives, methods, results, and next steps extracted in a single LLM call.
Pulls key findings, hypotheses, and conclusions with evidence strength scores from unstructured research logs.
Classifies research into scientific domain, subdomain, and keyword tags with a calibrated confidence score.
Compares two research logs side-by-side — surfacing shared themes, unique elements, and contradictions between studies.
Compiles multiple research summaries into an executive briefing with background, key findings, and actionable recommendations.
Semantic search across your research corpus — returns ranked excerpts with relevance scores for any concept or query.
A clean four-step pipeline — the model decides which tools to call, applies the optimized prompt variant, and returns typed JSON.
Upload research notes via the FastAPI endpoint or Streamlit dashboard. Persisted to the PostgreSQL-compatible database.
AgentCore sends the task to OpenAI with all 6 tools. The model decides which to call — zero manual routing logic.
PromptManager selects Variant A or B deterministically per session, enabling live A/B testing across all workflows.
Results are quality-scored, stored, and returned as typed JSON — available in the analytics dashboard or via REST.
Run any workflow in under a second. The agent handles tool selection, prompt optimization, and structured output automatically.
Every workflow runs two prompt variants simultaneously. Systematic iteration delivered a 38% quality improvement across all pipelines.
Chain-of-thought prompts with output constraints consistently outperform direct prompts across all 6 workflows.
Every layer chosen for reliability and extensibility — switch from SQLite to PostgreSQL with one env var change.
Type-annotated codebase with Pydantic v2 for runtime validation throughout.
Async REST API with OpenAPI docs auto-generated at /docs.
Native function calling API for agentic multi-tool orchestration — no LangChain abstractions.
Multi-page analytics dashboard with live charts, A/B results, and an agent chat interface.
SQLAlchemy ORM with full PostgreSQL support. SQLite used for zero-config local development.
Declarative ORM models with typed CRUD layer. Database created automatically on first run.
Deterministic variant assignment per session via MD5 hash — consistent UX, unbiased sampling.
python run.py starts the full stack. No Docker, no process manager required.
Clone, set your API key, and run the complete stack in under 3 minutes.