Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
56 commits
Select commit Hold shift + click to select a range
3b6685a
Add HOW_I_DID_IT.md for Explainable Knowledge Agent
praveen-singh-007 Apr 19, 2026
aa70df1
level-3: Praveen Singh
praveen-singh-007 Apr 19, 2026
caf1b8a
level-3: Praveen Singh
praveen-singh-007 Apr 19, 2026
ece26f2
level-3: Praveen Singh
praveen-singh-007 Apr 19, 2026
572933e
level-3: Praveen Singh
praveen-singh-007 Apr 19, 2026
02d8ccd
Revise HOW I DID IT for Level 2 submission
praveen-singh-007 Apr 19, 2026
1cc23dc
Revise Level 3 submission with evidence and examples
praveen-singh-007 Apr 19, 2026
a7514a5
Create llm_output.txt
praveen-singh-007 Apr 19, 2026
6d7b2ce
Add test output for LPI Sandbox
praveen-singh-007 Apr 19, 2026
c7c7ddf
Revise Level 3 documentation format
praveen-singh-007 Apr 19, 2026
a4ef174
Update level3.md
praveen-singh-007 Apr 19, 2026
2516093
Update level3.md with repo and code links
praveen-singh-007 Apr 19, 2026
995a7fb
Change repository link from lpi-life-agent to lpi-smile-agent
praveen-singh-007 Apr 19, 2026
9a48b50
Update level3.md
praveen-singh-007 Apr 19, 2026
7455cff
Update level3.md
praveen-singh-007 Apr 19, 2026
4e26039
Delete submissions/praveen-singh/level3 directory
praveen-singh-007 Apr 19, 2026
84e93b0
Create HOW_I_DID_IT.md
praveen-singh-007 Apr 19, 2026
884177b
Add Level 3 AI Agent documentation
praveen-singh-007 Apr 19, 2026
714394d
Create llm_output.txt
praveen-singh-007 Apr 19, 2026
32690f4
Delete submissions/praveen-singh/llm_output.txt
praveen-singh-007 Apr 19, 2026
33bfbb6
Add LLM output on SMILE methodology and reflections
praveen-singh-007 Apr 19, 2026
1155f69
Create test_output.txt
praveen-singh-007 Apr 19, 2026
a546d2e
Delete submissions/praveen-singh/level3 directory
praveen-singh-007 Apr 19, 2026
9b0f9ec
Add Level 3 submission for Explainable Knowledge Agent
praveen-singh-007 Apr 19, 2026
f75b52c
Delete submissions/praveen-singh/level3.md
praveen-singh-007 Apr 19, 2026
b8c040a
level-3: Praveen Singh
praveen-singh-007 Apr 19, 2026
a173943
level-3: Praveen Singh
praveen-singh-007 Apr 19, 2026
a4f5993
level-4: Praveen Singh
praveen-singh-007 Apr 19, 2026
50ab3f8
level-4: Praveen Singh
praveen-singh-007 Apr 19, 2026
a1f0811
level-4: Praveen Singh
praveen-singh-007 Apr 19, 2026
a432473
level-4: Praveen Singh
praveen-singh-007 Apr 19, 2026
308271a
level-4: Praveen Singh
praveen-singh-007 Apr 19, 2026
d382575
level-4: Praveen Singh
praveen-singh-007 Apr 19, 2026
22f4334
level-4: Praveen Singh
praveen-singh-007 Apr 19, 2026
e9689fa
level-4: Praveen Singh
praveen-singh-007 Apr 19, 2026
7f4acaf
level-3: Praveen Singh
praveen-singh-007 Apr 20, 2026
2792fe2
level-3: Praveen Singh
praveen-singh-007 Apr 20, 2026
e43417d
level-3: Praveen Singh
praveen-singh-007 Apr 20, 2026
14d8c76
level-3: Praveen Singh
praveen-singh-007 Apr 20, 2026
3753181
level-3: Praveen Singh
praveen-singh-007 Apr 20, 2026
9a7c5cd
level-3: Praveen Singh
praveen-singh-007 Apr 20, 2026
3a9c271
level-4: Praveen Singh
praveen-singh-007 Apr 20, 2026
94de445
level-4: Praveen Singh
praveen-singh-007 Apr 20, 2026
7239701
level-4: Praveen Singh
praveen-singh-007 Apr 20, 2026
e155c84
level-4: Praveen Singh
praveen-singh-007 Apr 20, 2026
f45058e
level-4: Praveen Singh
praveen-singh-007 Apr 20, 2026
b4b553b
level-4: Praveen Singh
praveen-singh-007 Apr 20, 2026
098c18c
level-4: Praveen Singh
praveen-singh-007 Apr 20, 2026
5dafa64
level-4: Praveen Singh
praveen-singh-007 Apr 20, 2026
d6d7cdb
level-4: Praveen Singh
praveen-singh-007 Apr 20, 2026
e74b80f
level-4: Praveen Singh
praveen-singh-007 Apr 20, 2026
f374b13
level-4: Praveen Singh
praveen-singh-007 Apr 20, 2026
9ee7f07
level-4: Praveen Singh
praveen-singh-007 Apr 20, 2026
d3495b0
level-4: Praveen Singh
praveen-singh-007 Apr 20, 2026
77df977
level-4: Praveen Singh
praveen-singh-007 Apr 20, 2026
49066a7
level-2: Praveen Singh
praveen-singh-007 Apr 21, 2026
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
150 changes: 150 additions & 0 deletions submissions/praveen-singh/HOW_I_DID_IT.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,150 @@
# How I Did It - Deployment Strategy Agent (Level 3)

## Approach

After Level 2, I decided to build a **decision-oriented agent** instead of a descriptive one.

The goal was simple:

```Given a use case and constraints, generate a realistic deployment strategy.```

I kept the implementation minimal and focused on:
- multi-tool reasoning
- constraint-aware output
- structured decision-making

---

## Key Decisions

### 1. Moving from explanation → decision agent

Instead of answering:
> “What are digital twins?”

I designed the agent to answer:
> “How should we build this under constraints?”

This required:
- structured outputs (architecture, risks, actions)
- justification of decisions

---

### 2. Expanding tool usage

Initial setup with 2 tools was not enough.

I moved to 4 tools:
- SMILE overview (methodology)
- Insights (scenario reasoning)
- Case studies (real grounding)
- Knowledge (context)

Decision:
> prioritize **multi-source grounding over simplicity**
---

### 3. Enforcing constraints as a core signal
Most LLM outputs ignore real-world limits.

I explicitly designed the agent to reason with:
- team size
- timeline
- infrastructure

Decision:
> treat constraints as **primary drivers**, not optional context"
---

### 4. Choosing structured output format
I forced the agent to always return:
- Architecture
- SMILE phases
- Risks
- What to avoid
- First actions
- Decision reasoning
Decision:
> structure improves both **quality and evaluation**
---
# Challenges Faced

### 1. MCP process failures
**Problem:**
- default: ValueError: I/O operation on closed file
- description: Reusing subprocess after `.communicate()`
- decision: spawn a new process per tool call
- outcome: Stable multi-tool execution

---

### 2. Weak reasoning from small model
**Problem:**
t-shallow outputs, generic answers
**Decision:** upgrade from `qwen2.5:1.5b` to `qwen2.5:7b`
**Outcome:**
- better structure
- improved reasoning
- fewer errors

---

### 3. Hallucinated technologies
**Problem:**
tool introduced tools/tech not in data; outputs looked impressive but incorrect.
**Decision:**
- explicitly block:
- invented technologies
- invented tools
enforce “use only provided data”
**Outcome:** More reliable outputs.

---

### 4. Irrelevant case study usage
**Problem:**
e.g., model used unrelated domains (e.g., heating systems for healthcare).
**Decision:**
- add relevance filtering:
- ignore cross-domain examples
- only use context-matching data.
**Outcome:** Improved correctness and credibility.

---

### 5. Over-engineered solutions
**Problem:**
del suggested complex systems despite tight constraints.
**Decision:**
- enforce:
- minimal viable twin (MVT)
- “simplest possible architecture”.
**Outcome:** Realistic, implementable strategies.

---

### 6. Prompt instability
**Problem:**
e.g., too strict → empty/generic output; too loose → hallucinations.
**Decision:**
balance:
- strict grounding rules
and flexible reasoning.
**Outcome:** Consistent, high-quality outputs.

## What I Learned
### 1. Prompt design > code
Most improvements came from:

highlighting:- refining instructions, enforcing constraints, guiding structure.
---
### 2. Constraints improve intelligence

Without constraints:
general answers.

With constraints:
appropriate, practical answers.
---
day-to-day learning about the importance of grounding and relevance in AI systems is crucial for reliable performance and trustworthy outputs.
62 changes: 62 additions & 0 deletions submissions/praveen-singh/level2.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,62 @@
# Level 2 Submission — Praveen Singh

## LPI Sandbox Setup

All 7 tools executed successfully, confirming that the LPI sandbox is functioning correctly. Using the test client felt like interacting with a modular system where each tool represents a specific capability of the agent. Instead of producing a single combined output, the system exposes well-defined functions, which clearly demonstrates how agents can operate through structured tool calls rather than relying entirely on raw LLM responses.

---

## Test Client Output

=== LPI Sandbox Test Client ===

[LPI Sandbox] Server started — 7 read-only tools available
Connected to LPI Sandbox

Available tools (7):

* smile_overview
* smile_phase_detail
* query_knowledge
* get_case_studies
* get_insights
* list_topics
* get_methodology_step

[PASS] smile_overview({})
[PASS] smile_phase_detail({"phase":"reality-emulation"})
[PASS] list_topics({})
[PASS] query_knowledge({"query":"explainable AI"})
[PASS] get_case_studies({})
[PASS] get_case_studies({"query":"smart buildings"})
[PASS] get_insights({"scenario":"personal health digital twin","tier":"free"})
[PASS] get_methodology_step({"phase":"concurrent-engineering"})

=== Results ===
Passed: 8/8
Failed: 0/8

All tools working. Your LPI Sandbox is ready.
You can now build agents that connect to this server.

---

## Local LLM Setup (Ollama)

**Model used:**
qwen2.5:1.5b

**Prompt:**
What is SMILE methodology?

**Response (summary):**
The model explained SMILE as a structured approach focused on managing the full lifecycle of information. It highlighted how processes like data creation, storage, access control, and deletion can be systematized, leading to better compliance, lower risk, and improved efficiency.

**Observation:**
Running the model locally felt noticeably different from using cloud APIs. Having direct control over execution made the process more transparent and gave it a system-level feel, rather than just sending queries to an external service.

---

## Reflection on SMILE Methodology

SMILE comes across more as a systems engineering approach than just a standard methodology. A key takeaway for me was its focus on designing systems that enforce correct behavior by default, instead of depending on manual rules or user discipline. This aligns closely with how scalable AI systems should be built, where reliability is embedded into the architecture itself. It also connects naturally with digital twins, where continuous data flow and lifecycle awareness are essential for generating meaningful insights. Overall, it shifts the perspective from simply building models to understanding and designing how systems evolve and operate over time.
Loading
Loading