This notebook demonstrates how to build a Multi-agent workflow by integrating LangChain with LangGraph Studio, allowing you to orchestrate multiple specialized agents for gathering, analyzing, and synthesizing information. In this tutorial, we focus on researching a specific person, their professional background, and the company they work for, as well as generating relevant follow-up questions or interview prompts.
By visualizing this agent workflow in LangGraph Studio, you can easily debug, modify, and extend the pipeline. Each agent’s output can be inspected step by step, making it straightforward to add new components or adjust the process flow.
You can alternatively set API keys such as ANTHROPIC_API_KEY in a .env file and load them.
[Note] This is not necessary if you've already set the required API keys in previous steps.
# Load API keys from .env file
from dotenv import load_dotenv
load_dotenv(override=True)
True
What is LangGraph Studio
LangGraph Studio offers a new way to develop LLM applications by providing a specialized agent IDE that enables visualization, interaction, and debugging of complex agentic applications.
With visual graphs and the ability to edit the state, you can better understand agent workflows and iterate faster. LangGraph Studio integrates with LangSmith so you can collaborate with teammates to debug failure modes.
# Get tavily api key
from tavily import AsyncTavilyClient
# Search
tavily_async_client = AsyncTavilyClient()
from dataclasses import dataclass, field
from typing import Any, Optional, Annotated
from pydantic import BaseModel, Field
import operator
from pydantic import BaseModel
class Person(BaseModel):
"""A class representing a person to research."""
name: Optional[str] = None
"""The name of the person."""
company: Optional[str] = None
"""The current company of the person."""
linkedin: Optional[str] = None
"""The Linkedin URL of the person."""
email: str
"""The email of the person."""
role: Optional[str] = None
"""The current title of the person."""
@dataclass(kw_only=True)
class InputState:
"""Input state defines the interface between the graph and the user (external API)."""
person: Person
"Person to research."
user_notes: Optional[dict[str, Any]] = field(default=None)
"Any notes from the user to start the research process."
@dataclass(kw_only=True)
class OverallState:
"""Input state defines the interface between the graph and the user (external API)."""
person: Person
"Person to research provided by the user."
user_notes: str = field(default=None)
"Any notes from the user to start the research process."
search_queries: list[str] = field(default=None)
"List of generated search queries to find relevant information"
# Add default values for required fields
completed_notes: Annotated[list, operator.add] = field(default_factory=list)
"Notes from completed research related to the schema"
info: dict[str, Any] = field(default=None)
"""
A dictionary containing the extracted and processed information
based on the user's query and the graph's execution.
This is the primary output of the enrichment process.
"""
Questions: list[str] = field(default=None)
is_satisfactory: bool = field(default=None)
"True if all required fields are well populated, False otherwise"
combined_notes: str = field(default_factory=str)
"Consolidated research notes combining all gathered information into a single coherent document."
project_queries: list[str] = field(default_factory=list)
"List of search queries specifically focused on finding project-related information."
company_search_queries: list[str] = field(default=None)
"List of search queries generated for gathering company-specific information."
reflection_steps_taken: int = field(default=0)
"Number of times the reflection node has been executed"
@dataclass(kw_only=True)
class OutputState:
"""The response object for the end user.
This class defines the structure of the output that will be provided
to the user after the graph's execution is complete.
"""
info: dict[str, Any]
"""
A dictionary containing the extracted and processed information
based on the user's query and the graph's execution.
This is the primary output of the enrichment process.
"""
class Queries(BaseModel):
queries: list[str] = Field(
description="List of search queries.",
)
class ReflectionOutput(BaseModel):
is_satisfactory: bool = Field(
description="True if all required fields are well populated, False otherwise"
)
search_queries: list[str] = Field(
description="If is_satisfactory is False, provide 1-3 targeted search queries to find the missing information"
)
reasoning: str = Field(description="Brief explanation of the assessment")
def deduplicate_and_format_sources(
search_response, max_tokens_per_source, include_raw_content=True
):
"""
Takes either a single search response or list of responses from Tavily API and formats them.
Limits the raw_content to approximately max_tokens_per_source.
include_raw_content specifies whether to include the raw_content from Tavily in the formatted string.
Args:
search_response: Either:
- A dict with a 'results' key containing a list of search results
- A list of dicts, each containing search results
Returns:
str: Formatted string with deduplicated sources
"""
# Convert input to list of results
if isinstance(search_response, dict):
sources_list = search_response["results"]
elif isinstance(search_response, list):
sources_list = []
for response in search_response:
if isinstance(response, dict) and "results" in response:
sources_list.extend(response["results"])
else:
sources_list.extend(response)
else:
raise ValueError(
"Input must be either a dict with 'results' or a list of search results"
)
# Deduplicate by URL
unique_sources = {}
for source in sources_list:
if source["url"] not in unique_sources:
unique_sources[source["url"]] = source
# Format output
formatted_text = "Sources:\n\n"
for i, source in enumerate(unique_sources.values(), 1):
formatted_text += f"Source {source['title']}:\n===\n"
formatted_text += f"URL: {source['url']}\n===\n"
formatted_text += (
f"Most relevant content from source: {source['content']}\n===\n"
)
if include_raw_content:
# Using rough estimate of 4 characters per token
char_limit = max_tokens_per_source * 4
# Handle None raw_content
raw_content = source.get("raw_content", "")
if raw_content is None:
raw_content = ""
if len(raw_content) > char_limit:
raw_content = raw_content[:char_limit] + "... [truncated]"
formatted_text += f"Full source content limited to {max_tokens_per_source} tokens: {raw_content}\n\n"
return formatted_text.strip()
import os
from dataclasses import dataclass, fields
from typing import Any, Optional
from langchain_core.runnables import RunnableConfig
@dataclass(kw_only=True)
class Configuration:
max_search_queries: int = 3
max_search_results: int = 3
max_reflection_steps: int = 0
@classmethod
def from_runnable_config(cls, config: Optional[RunnableConfig] = None) -> "Configuration":
configurable = config["configurable"] if config and "configurable" in config else {}
values: dict[str, Any] = {
"max_search_queries": configurable.get("max_search_queries", 3),
"max_search_results": configurable.get("max_search_results", 3),
"max_reflection_steps": configurable.get("max_reflection_steps", 0),
}
return cls(**values)
QUERY_WRITER_PROMPT = """ You are a search query generator tasked with creating targeted search queries to gather specific information about a person.
Here is the person you are researching: {person}
Generate at most {max_search_queries} search queries that will help gather the following information:
<user_notes>
{user_notes}
</user_notes>
Your query should:
1. Make sure to look up the right name
2. Use context clues as to the company the person works at (if it isn't concretely provided)
3. Do not hallucinate search terms that will make you miss the persons profile entirely
4. Take advantage of the Linkedin URL if it exists, you can include the raw URL in your search query as that will lead you to the correct page guaranteed.
Create a focused query that will maximize the chances of finding schema-relevant information about the person.
Remember we are interested in determining their work experience mainly."""
# generate queries
from langchain_core.runnables import RunnableConfig
from typing import cast
def generate_queries(state: OverallState, config: RunnableConfig) -> dict[str, Any]:
"""Generate search queries based on the user input and extraction schema."""
# Get configuration
configurable = Configuration.from_runnable_config(config)
max_search_queries = configurable.max_search_queries
# Generate search queries
structured_llm = llm.with_structured_output(Queries)
# Format system instructions
person_str = f"Email: {state.person['email']}"
if "name" in state.person:
person_str += f" Name: {state.person['name']}"
if "linkedin" in state.person:
person_str += f" LinkedIn URL: {state.person['linkedin']}"
if "role" in state.person:
person_str += f" Role: {state.person['role']}"
if "company" in state.person:
person_str += f" Company: {state.person['company']}"
query_instructions = QUERY_WRITER_PROMPT.format(
person=person_str,
user_notes=state.user_notes,
max_search_queries=max_search_queries,
)
# Generate queries
results = cast(
Queries,
structured_llm.invoke(
[
{"role": "system", "content": query_instructions},
{
"role": "user",
"content": "Please generate a list of search queries related to the schema that you want to populate.",
},
]
),
)
# Queries
query_list = [query for query in results.queries]
return {"search_queries": query_list}
SEARCH_COMPANY_PROMPT = """You are a search query generator tasked with creating targeted search queries to gather specific information about a company.
Person and Company Information: {person}
Additional Context:
{user_notes}
Your query should:
1. Focus on finding detailed company information and recent developments
2. Look for company products, services, and main business areas
3. Search for company culture, work environment, and growth trajectory
4. Include specific company name variations and common abbreviations if known
Constraints:
1. Generate at most {max_search_queries} unique and meaningful queries
2. Do not use generic terms that might dilute search results
3. Prioritize recent company information (within last 2 years)
4. Include company name in each query for better targeting
Create focused queries that will maximize the chances of finding comprehensive company information.
Remember we are interested in understanding the company's business context, culture, and market position.
"""
import asyncio
async def generate_queries_for_company(state: OverallState, config: RunnableConfig) -> dict[str, Any]:
"""
Generate search queries specifically for the company,
possibly merging with user-provided queries (company_queries in state).
"""
configurable = Configuration.from_runnable_config(config)
max_search_queries = configurable.max_search_queries
person_str = f"Email: {state.person['email']}"
if "name" in state.person:
person_str += f" Name: {state.person['name']}"
if "linkedin" in state.person:
person_str += f" LinkedIn URL: {state.person['linkedin']}"
if "role" in state.person:
person_str += f" Role: {state.person['role']}"
if "company" in state.person:
person_str += f" Company: {state.person['company']}"
structured_llm = llm.with_structured_output(Queries) # Queries: pydantic model for array of strings
prompt = SEARCH_COMPANY_PROMPT.format(
person=person_str,
user_notes=state.user_notes,
max_search_queries=max_search_queries,
)
llm_result = cast(
Queries,
await structured_llm.ainvoke(
[
{"role": "system", "content": prompt},
{
"role": "user",
"content": "Generate company-related search queries now.",
},
]
)
)
generated_queries = [q for q in llm_result.queries]
return {"company_search_queries": generated_queries}
COMPANY_INFO_PROMPT = """You are doing web research on company.
The following schema shows the type of information we're interested in:
You have just scraped website content. Your task is to take clear, organized notes about a company, focusing on topics relevant to our interests.
<Website contents>
{content}
</Website contents>
Here are any additional notes from the user:
<user_notes>
{user_notes}
</user_notes>
Please provide detailed research notes that:
1. Are well-organized and easy to read
2. Include specific facts, dates, and figures when available
3. Maintain accuracy of the original content
4. Note when important information appears to be missing or unclear
Remember: Just take clear notes that capture all relevant information."""
async def research_company(state: OverallState, config: RunnableConfig) -> dict[str, Any]:
"""
Use 'company_search_queries' to search relevant info about the company,
then produce 'company_notes'.
"""
configurable = Configuration.from_runnable_config(config)
max_search_results = configurable.max_search_results
# Web search
search_tasks = []
for query in state.company_search_queries:
search_tasks.append(
tavily_async_client.search(
query,
days=360,
max_results=max_search_results,
include_raw_content=True,
topic="general",
)
)
# Execute all searches concurrently
search_docs = await asyncio.gather(*search_tasks)
# Deduplicate and format sources
source_str = deduplicate_and_format_sources(
search_docs, max_tokens_per_source=1000, include_raw_content=True
)
p = COMPANY_INFO_PROMPT.format(
content=source_str,
user_notes=state.user_notes,
)
result = await llm.ainvoke(p)
return {"company_notes": str(result.content)}
INFO_PROMPT = """You are doing web research on people, {people}.
The following schema shows the type of information we're interested in:
You have just scraped website content. Your task is to take clear, organized notes about a person, focusing on topics relevant to our interests.
<Website contents>
{content}
</Website contents>
Here are any additional notes from the user:
<user_notes>
{user_notes}
</user_notes>
Please provide detailed research notes that:
1. Are well-organized and easy to read
2. Focus on topics mentioned in the schema
3. Include specific facts, dates, and figures when available
4. Maintain accuracy of the original content
5. Note when important information appears to be missing or unclear
Remember: Don't try to format the output to match the schema - just take clear notes that capture all relevant information."""
async def research_person(state: OverallState, config: RunnableConfig) -> dict[str, Any]:
"""Execute a multi-step web search and information extraction process.
This function performs the following steps:
1. Executes concurrent web searches using the Tavily API
2. Deduplicates and formats the search results
"""
# Get configuration
configurable = Configuration.from_runnable_config(config)
max_search_results = configurable.max_search_results
# Web search
search_tasks = []
for query in state.search_queries:
search_tasks.append(
tavily_async_client.search(
query,
days=360,
max_results=max_search_results,
include_raw_content=True,
topic="general",
)
)
# Execute all searches concurrently
search_docs = await asyncio.gather(*search_tasks)
# Deduplicate and format sources
source_str = deduplicate_and_format_sources(
search_docs, max_tokens_per_source=1000, include_raw_content=True
)
# Generate structured notes relevant to the extraction schema
p = INFO_PROMPT.format(
content=source_str,
people=state.person,
user_notes=state.user_notes,
)
result = await llm.ainvoke(p)
return {"completed_notes": [str(result.content)]}
SEARCH_PROJECTS_PROMPT = """You have the following notes about a person:
{completed_notes}
From these notes, extract up to 3 specific search queries that would help us find more detailed information
about projects, case studies, or portfolio items mentioned or implied in the notes.
Generate at most {max_search_queries} search queries that will help gather the following information:
Output them as JSON in the format: {{"queries": ["query1", "query2", ...]}}.
"""
async def extract_project_queries(state: OverallState, config: RunnableConfig) -> dict[str, Any]:
"""
Examine `state.completed_notes` to identify potential project-based keywords or queries,
then return them as a list so we can further search Tavily.
"""
configurable = Configuration.from_runnable_config(config)
max_search_queries = configurable.max_search_queries
# 2) We'll ask LLM to parse out project-related queries
structured_llm = llm.with_structured_output(Queries) # Re-using your Queries pydantic model
prompt = SEARCH_PROJECTS_PROMPT.format(
completed_notes = state.completed_notes,
max_search_queries = max_search_queries,
)
llm_result = await structured_llm.ainvoke(
[
{"role": "system", "content": prompt},
{"role": "user", "content": "Please provide the project-related queries now."},
]
)
# 3) Return the queries
return {"project_queries": llm_result.queries if llm_result else []}
PROJECT_INFO_PROMPT = """You are doing additional web research on projects by {people}.
Focus on details about important projects they have worked on. Gather:
- project name
- relevant context or goals
- timeline
- major accomplishments
- technologies used
- challenges faced
Website contents:
{content}
Return your notes in a well-structured format that could be used for interview preparation.
"""
async def research_projects(state: OverallState, config: RunnableConfig) -> dict[str, Any]:
"""
Execute a multi-step web search focusing on the person's projects,
guided by the project queries extracted from `completed_notes`.
"""
# 1) Retrieve the queries from state
project_queries = getattr(state, "project_queries", [])
if not project_queries:
return {"project_notes": "No project queries found. Please run extract_project_queries first."}
# 2) Set up concurrency
configurable = Configuration.from_runnable_config(config)
max_search_results = configurable.max_search_results
search_tasks = []
for query in project_queries:
search_tasks.append(
tavily_async_client.search(
query,
days=360,
max_results=max_search_results,
include_raw_content=True,
topic="general",
)
)
# 3) Execute all searches concurrently
search_results = await asyncio.gather(*search_tasks)
# 4) Format & deduplicate
source_str = deduplicate_and_format_sources(
search_results, max_tokens_per_source=1000, include_raw_content=True
)
# 5) Summarize the project details with LLM
prompt = PROJECT_INFO_PROMPT.format(
people=state.person,
content=source_str,
)
llm_result = await llm.ainvoke(prompt)
return {"project_notes": str(llm_result.content)}
COMBINE_ALL_PROMPT = """We have three sets of notes:
1) Person Notes:
{person_notes}
2) Company Notes:
{company_notes}
3) Project Notes:
{project_notes}
Please merge these notes into a single coherent summary.
Remove any duplicate or conflicting information.
The result should flow logically, covering person's details, the company background, and key project info.
"""
async def combine_notes(state: OverallState) -> dict[str, Any]:
"""
Merge the notes from research_person (person_notes) and research_projects (project_notes)
into a single cohesive note.
"""
person_notes = getattr(state, "completed_notes", "")
company_notes = getattr(state, "company_notes", "")
project_notes = getattr(state, "project_notes", "")
prompt = COMBINE_ALL_PROMPT.format(
person_notes=person_notes,
company_notes=company_notes,
project_notes=project_notes,
)
llm_result = await llm.ainvoke(prompt)
return {"combined_notes": str(llm_result.content)}
GENERATE_QUESTIONS_PROMPT = """Based on the following combined note:
{combined_notes}
Create a list of interview questions that focus on this person's experiences,
skills, and the project/product details mentioned. Format them as:
Q1: ...
Q2: ...
Q3: ...
"""
async def generate_questions(state: OverallState) -> dict[str, Any]:
"""
Use the combined notes to generate a set of interview questions.
"""
combined_notes = getattr(state, "combined_notes", "")
if not combined_notes:
return {"Questions": "No combined notes available. Please run combine_notes first."}
prompt = GENERATE_QUESTIONS_PROMPT.format(
combined_notes=combined_notes
)
llm_result = await llm.ainvoke(prompt)
return {"Questions": str(llm_result.content)}
REFLECTION_PROMPT = """
You are verifying if the following interview question is reasonably supported by the combined notes.
If the question is even partially related to or inspired by the notes, set `is_satisfactory` to True.
Only set `is_satisfactory` to False if it clearly contradicts or has no logical connection.
<Question>
{Question}
</Question>
<combined_notes>
{combine_notes}
</combined_notes>
Important:
- If any part of the combined notes is relevant to this question, consider it supported.
- Only answer with 'true' or 'false' as is_satisfactory. Provide a brief explanation.
"""
def reflection(state: OverallState) -> dict[str, Any]:
"""
Analyze the quality and completeness of the gathered information.
"""
structured_llm = llm.with_structured_output(ReflectionOutput)
system_prompt = REFLECTION_PROMPT.format(
Question=state.Questions,
combine_notes=state.combined_notes
)
result = structured_llm.invoke(
[
{"role": "system", "content": system_prompt},
{"role": "user", "content": "Produce a structured reflection output."},
]
)
result = cast(ReflectionOutput, result)
if result.is_satisfactory:
return {"is_satisfactory": result.is_satisfactory}
else:
return {
"is_satisfactory": result.is_satisfactory,
"reflection_steps_taken": state.reflection_steps_taken + 1,
"reasoning": result.reasoning
}
from typing import Literal
from langgraph.graph import END
def route_from_reflection(
state: OverallState, config: RunnableConfig
) -> Literal[END, "generate_questions"]: # type: ignore
"""Route the graph based on the reflection output."""
configurable = Configuration.from_runnable_config(config)
# If we have satisfactory results, end the process
if state.is_satisfactory:
return END
# If results aren't satisfactory but we haven't hit max steps, continue research
if state.reflection_steps_taken <= configurable.max_reflection_steps:
return "generate_questions"
return END
from langchain_opentutorial.graphs import visualize_graph
visualize_graph(graph)
import asyncio
from typing import Any, Dict, List, Callable
from collections import defaultdict
from langchain_core.runnables import RunnableConfig
from langgraph.graph.state import CompiledStateGraph
async def astream_graph(
graph: CompiledStateGraph,
inputs: Dict[str, Any],
config: RunnableConfig,
node_names: List[str] = [],
callback: Callable[[Dict[str, str]], None] = None,
):
"""
This function executes a compiled LangGraph and collects the output chunks for each node,
storing them in memory. After the entire graph execution finishes, it prints the results
for each node in the order they first appeared.
**Key Point:**
- This approach **disables real-time streaming** (no partial outputs to the console as they arrive).
Instead, all partial outputs are aggregated and printed only after graph execution is complete.
This helps to avoid interleaved or mixed outputs when different nodes run in parallel.
**Note:** If you run the same graph in **LangGraph Studio**, you can observe real-time chunk streaming.
However, with this function, we intentionally buffer all chunks to produce a cleaner final output
for each node.
"""
# Nodes to exclude from output
excluded_nodes = {
"generate_queries",
"generate_queries_for_company",
"extract_project_queries",
"reflection"
}
# Keep track of the order in which nodes appear
node_order: List[str] = []
# A dictionary to store buffered text for each node
node_buffers: Dict[str, str] = defaultdict(str)
# Consume the asynchronous stream of chunks
async for chunk_msg, metadata in graph.astream(inputs, config, stream_mode="messages"):
curr_node = metadata["langgraph_node"]
# If node_names is specified and curr_node is not in it, skip processing
if node_names and curr_node not in node_names:
continue
# If this node is excluded, skip
if curr_node in excluded_nodes:
continue
# If callback is provided, you could handle partial chunks in real time.
# But here, we want to buffer everything until completion, so we skip.
if curr_node not in node_order:
node_order.append(curr_node)
# Accumulate chunks in the buffer for this node
node_buffers[curr_node] += chunk_msg.content
# After the entire graph finishes, output results for each node in the order they first appeared
for node in node_order:
# If the node is excluded, skip it
if node in excluded_nodes:
continue
# Print final aggregated content for this node
print("\n" + "=" * 50)
print(f"🔄 Node: \033[1;36m{node}\033[0m 🔄")
print("- " * 25)
print(node_buffers[node], end="", flush=True)
config = RunnableConfig(
recursion_limit=10,
configurable={
"max_search_queries": 3,
"max_search_results": 3,
"max_reflection_steps": 1,
},
)
input_data = {
"person": {
"email": "jhkim@reversemountain.co.kr",
}}
# This function buffers all partial outputs for each node and prints them only after the entire graph execution completes.
# Hence, if the graph takes a while to run, say 40-50 seconds, you won't see any intermediate output until it finishes!
await astream_graph(graph, input_data, config)
==================================================
🔄 Node: research_company 🔄
- - - - - - - - - - - - - - - - - - - - - - - - -
RESEARCH NOTES: REVERSE MOUNTAIN / TICKI TACKA
COMPANY OVERVIEW
- Company Name: Reverse Mountain Corp (리버스마운틴)
- Founded: August 2021
- Location: Seoul, South Korea (Jongno Youth Startup Center)
- CEO: Gyeongmin KIM
- Current Employees: 9 (as of 2024)
CORE PRODUCT
- Product Name: Ticki Tacka
- Type: AI-powered enterprise project control and performance optimization system
- Key Features:
* Cascading goals system (team objectives to weekly KPIs)
* Transparent contribution tracking
* Weekly evaluation and feedback system
* Focus on workplace engagement for 20-30 age demographic
COMPANY HISTORY
- Initially started with two hobby platforms (Palette and Hobby Picker)
- Pivoted to workplace solutions after identifying workplace pain points
- Current focus on enterprise performance management solutions
FUNDING & ACHIEVEMENTS
- Investment Stage: Seed
- Recent Developments:
* Received TIPS (Tech Incubator Program for Startup) selection
* Secured seed investment from CompanyX
* Won Grand Prize at Sungkyunkwan University Demo Day
* Selected for Global Accelerating Program
- Currently in implementation discussions with 3 large U.S. companies (reported $1T+ revenue)
NOTABLE PARTNERSHIPS/RECOGNITION
- Selected for TIPS R&D funding (up to 500M won for 2 years)
- Additional funding for business development and overseas marketing (up to 100M won each)
- CompanyX accelerator participation
MISSING/UNCLEAR INFORMATION
- Specific revenue figures
- Total funding amount
- Detailed product pricing
- Market share/user numbers
- International presence details beyond U.S. discussions
- Complete list of investors
Note: Information compiled from multiple Korean sources and LinkedIn profile of CEO. Some details may require verification due to translation and limited public information.
==================================================
🔄 Node: research_person 🔄
- - - - - - - - - - - - - - - - - - - - - - - - -
RESEARCH NOTES - Jihyun Kim (jhkim@reversemountain.co.kr)
CURRENT ROLE
- AI Engineer at Reversemountain Corp
- Started August 2024 - Present
- Focus: NLP | LLM RAG
- Location: Seoul, South Korea
PREVIOUS EXPERIENCE
- Data Analyst at Tiffany & Co.
- November 2019 - November 2023
- Role involved sales data analysis
EDUCATION
- Currently enrolled at Korea National Open University
- Statistics and Data Science major
- Started August 2023
- Degree completion status unclear
SKILLS & EXPERTISE
- Specializes in AI, particularly LLM (Large Language Models) and RAG
- Has expressed specific interest in LLM technology
- Language skills:
- English (Limited working proficiency)
- Chinese (Full professional proficiency)
COMPANY CONTEXT
Reversemountain Corp:
- Founded: August 19, 2021
- CEO: Gyeongmin Kim
- Location: 2nd floor, Orange Planet, 217 Teheran-ro, Gangnam-gu, Seoul
- Employee count: 8 persons
- Main product: Ticki tacka (AI-powered enterprise project control and performance optimization system)
MISSING/UNCLEAR INFORMATION
- Specific educational background details beyond current enrollment
- Previous roles or positions between Tiffany & Co. and Reversemountain
- Technical certifications or specialized training
- Specific achievements or projects in current role
- Detailed technical skills beyond general AI/LLM focus
NOTE: The email provided matches the company domain (reversemountain.co.kr) where the subject is currently employed as an AI Engineer.
==================================================
🔄 Node: research_projects 🔄
- - - - - - - - - - - - - - - - - - - - - - - - -
Based on the available sources, here are the key project details for the person associated with jhkim@reversemountain.co.kr:
PROJECT: Ticki Tacka
Company: Reversemountain Corp
Timeline: 2021-Present
Description:
- AI-powered enterprise project control and performance optimization system
- Focused on improving workplace happiness and engagement, particularly for younger generations (20-30 age group)
Key Features:
- Cascading goals system that breaks team objectives into weekly KPIs
- Transparent contribution tracking showing each team member's impact
- Regular feedback through weekly evaluations
- AI-powered meeting optimization:
- Pre-meeting agenda and reference sharing
- Template-based meeting minutes
- Automatic recording and summarization using STT (Speech-to-Text) AI
- Action plan assignment and tracking
Technologies:
- AI/Machine Learning
- Speech-to-Text
- LLM (Large Language Models)
- RAG (Retrieval-Augmented Generation)
Challenges & Solutions:
- Initial failed attempts with hobby platforms (Palette and Hobby Picker)
- Pivoted after identifying workplace pain as a key issue
- Recognized that traditional surveillance/force approaches don't work with younger employees
- Developed system focused on showing benefits of engagement
Results/Impact:
- Successfully launched and operating with 11-50 employees
- Provides systematic solution for workplace engagement vs. individual manager efforts
- Helps companies improve employee happiness through structured goal-setting and recognition
Context:
- Company founded with mission to improve people's happiness levels
- Developed in response to research showing declining happiness as people age
- Aims to create systemic changes in workplace culture and engagement
- Plans to expand beyond work into other life aspects (health, relationships, personal development)
This appears to be the main significant project associated with this email, as it represents the core product of Reversemountain Corp.
==================================================
🔄 Node: combine_notes 🔄
- - - - - - - - - - - - - - - - - - - - - - - - -
Based on the available notes (only Person Notes provided), here's a coherent summary:
Professional Summary - Jihyun Kim
--------------------------------
Current Position:
- AI Engineer at Reversemountain Corp (August 2024 - Present)
- Focus areas: NLP and LLM RAG
- Location: Seoul, South Korea
- Contact: jhkim@reversemountain.co.kr
Professional Background:
- Previous role: Data Analyst at Tiffany & Co. (November 2019 - November 2023)
- Specialized in sales data analysis
Education:
- Currently pursuing Statistics and Data Science
- Korea National Open University (Started August 2023)
Technical Expertise:
- Specialization in AI, particularly LLM and RAG systems
- Language Proficiency:
- Chinese (Full professional proficiency)
- English (Limited working proficiency)
Company Context - Reversemountain Corp:
- Founded: August 19, 2021
- CEO: Gyeongmin Kim
- Location: 2nd floor, Orange Planet, 217 Teheran-ro, Gangnam-gu, Seoul
- Size: 8 employees
- Main Product: Ticki tacka (AI-powered enterprise project control and performance optimization system)
Note: Since only Person Notes were provided, this summary focuses on available information about Jihyun Kim and their current employer. Additional details about company operations or specific projects would require the missing Company and Project Notes.
==================================================
🔄 Node: generate_questions 🔄
- - - - - - - - - - - - - - - - - - - - - - - - -
Here are relevant interview questions based on the provided information:
Q1: Could you describe your transition from data analysis at Tiffany & Co. to AI engineering at Reversemountain Corp? What motivated this career shift?
Q2: How does your experience with sales data analysis at Tiffany & Co. contribute to your current work with NLP and LLM RAG systems?
Q3: Can you explain your role in developing or maintaining Ticki tacka, and what specific AI components you work with in this enterprise project control system?
Q4: How do you balance your current full-time role as an AI Engineer with pursuing your Statistics and Data Science degree? What synergies have you found between your studies and work?
Q5: Given your language proficiencies in Chinese and English, how do you utilize these skills in your current role, particularly in developing multilingual AI solutions?
Q6: As one of eight employees in a relatively young company, what challenges and opportunities have you encountered working in such a small team environment?
Q7: Could you elaborate on your specific experience with RAG (Retrieval-Augmented Generation) systems and how you're applying this technology at Reversemountain?
Q8: What metrics or KPIs do you use to measure the success of your AI implementations in Ticki tacka?
Q9: How has your background in sales data analysis influenced your approach to developing AI solutions for enterprise project control?
Q10: Being part of a startup founded in 2021, what role have you played in shaping the company's AI strategy and technical direction?
Jupyter Notebook Code Cell Extractor
This script converts Jupyter Notebook cells into a Python script with the following features:
Converts pip install magic commands into executable Python code
Removes or comments out visualization-related code
Handles cell deduplication
Processes cells up to the graph compilation
Maintains code organization and readability
This conversion is necessary because LangGraph Studio requires Python (.py) files for execution. This script helps transform our tutorial notebook into the correct format while maintaining all functionality.
Key Features:
Automatic package installation code generation
Cell content deduplication
Selective cell processing
Magic command handling
Proper formatting for LangGraph Studio compatibility
import json
import os
def extract_code_cells(notebook_path, output_path):
os.makedirs(os.path.dirname(output_path), exist_ok=True)
with open(notebook_path, 'r', encoding='utf-8') as f:
notebook = json.load(f)
def transform_cell(cell_source):
"""Transform Jupyter notebook code cell content"""
if isinstance(cell_source, list):
cell_source = ''.join(cell_source)
# Remove %%capture magic
if cell_source.startswith('%%capture'):
cell_source = '\n'.join(cell_source.split('\n')[1:])
# Transform %pip install commands to executable code
if cell_source.strip().startswith('%pip install'):
packages = [pkg.strip() for pkg in cell_source.replace('%pip install', '').strip().split()]
if not packages:
return ""
# Remove duplicates
packages = list(dict.fromkeys(packages))
return f'''import sys
import subprocess
def pip_install(package: str):
subprocess.check_call([sys.executable, "-m", "pip", "install", package])
if __name__ == "__main__":
packages = {packages}
for pkg in packages:
pip_install(pkg)
'''
# Handle visualization code
lines = cell_source.split('\n')
new_lines = []
for line in lines:
if 'from langchain_opentutorial.graphs import visualize_graph' in line \
or 'visualize_graph(graph)' in line:
new_lines.append('# ' + line)
else:
new_lines.append(line)
return '\n'.join(new_lines)
# Prevent duplicate cells
seen_cells = set()
unique_cells = []
found_compile = False
# Process notebook cells
for cell in notebook['cells']:
if cell['cell_type'] == 'code':
cell_content = ''.join(cell['source'])
# Set flag when compile is found
if 'graph = builder.compile()' in cell_content:
found_compile = True
# Handle duplicates
if cell_content not in seen_cells:
seen_cells.add(cell_content)
transformed = transform_cell(cell['source'])
if transformed.strip():
unique_cells.append(transformed)
# Stop processing after compile
if found_compile:
break
# Save transformed code
with open(output_path, 'w', encoding='utf-8') as f:
for cell in unique_cells:
f.write(cell + '\n\n')
# Usage example
notebook_file = "./20-LangGraphStudio-MultiAgent.ipynb"
output_file = "./langgraph_studio/output_script.py"
extract_code_cells(notebook_file, output_file)
How to connect a local agent to LangGraph Studio
Connection Options
There are two ways to connect your local agent to LangGraph Studio:
Currently, the desktop application only supports only macOS. Other users can run a local LangGraph server and use the web studio. We also depend on Docker Engine to be running. Currently, we support only the following runtimes:
First, you will need to setup your application in the proper format. This means defining a langgraph.json file which contains paths to your agent(s). See this guide for information on how to do so.
Please make sure that all the required files for running LangGraph Studio are located in the langgraph_studio folder.
For this example, we will use this example repository here which uses a requirements.txt file for dependencies:
As previously mentioned, we are using Docker Desktop , so please download it, launch the app, and make sure the Docker engine is running. Then, in LangGraph Studio , open the langgraph_studio folder.
Demo
Here is a demo video demonstrating how it works in practice.
Here are relevant interview questions based on the provided notes:
Technical Experience & Skills:
Q1: Could you describe your transition from data analysis at Tiffany & Co. to AI engineering, and how your previous experience informs your current work with LLMs and RAG systems?
Q2: What specific NLP challenges have you encountered while developing the "Ticki tacka" project control system, and how have you addressed them?
Q3: How do you balance your current studies in Statistics and Data Science with your role as an AI Engineer? What aspects of your coursework directly apply to your work?
Project & Product Specific:
Q4: Could you walk us through the core AI components of the "Ticki tacka" system and your role in its development?
Q5: What metrics or KPIs have you established to measure the effectiveness of the AI-powered workplace optimization solutions you're developing?
Language & Communication:
Q6: Given your Chinese and English language proficiencies, how do you leverage these skills in your current role, particularly in technical documentation or team collaboration?
Company Growth & Vision:
Q7: How has the recent seed funding and TIPS grant influenced your team's approach to AI development and project priorities?
Q8: What role do you see AI playing in workplace happiness and employee engagement, and how does this align with Reversemountain's mission?
Technical Implementation:
Q9: Could you describe your experience implementing RAG systems, and what challenges have you encountered in enterprise applications?
Q10: How do you approach the balance between model performance and practical business requirements in your AI solutions?
Wow, these interview questions are really well-tailored based on my past and current companies!
I’d better make sure I don’t get caught off guard if they actually come up in an interview. 🤣
After a short while, once the build completes successfully, you will see a screen similar to the one below.
Now, let’s run a test. I’ll enter my actual company email address.