Goal: Build a language model system (≤1B parameters) to generate semiconductor device simulation code from natural language circuit design specifications, and design your own benchmark to evaluate model performance.
About the Simulation Platform: This project uses Silvaco TCAD (Technology Computer-Aided Design), an industry-standard semiconductor simulation platform for device modeling and circuit analysis. You will train models to generate SPICE-compatible simulation code that describes semiconductor device structures and electrical characteristics.
Possible Approaches: You can explore various techniques such as fine-tuning (LoRA, QLoRA), prompt engineering, retrieval-augmented generation (RAG), chain-of-thought prompting, or any combination that works best for your solution. The choice of methodology is completely open.
Benchmark Design New: You are responsible for designing a comprehensive benchmark to evaluate your model's code generation capabilities. This includes creating test cases, defining evaluation metrics, and demonstrating rigorous assessment of your model's strengths and weaknesses.
Format: Individual or team (2-3 students)
Final Presentation: December 3 (10 minutes per team)
Final Submission: December 12, 11:59 PM
Download:
Download Dataset (24 MB)Contents:
silvaco_dataset_train.json - 713 instruction-code pairs for trainingSilvaco_Examples_Student.zip - 726 reference .in files + 76 .lib filesREADME.md - Complete dataset documentationImportant: You will design your own benchmark to evaluate your model. Focus on creating diverse, challenging test cases that assess generalization, not memorization.
Dataset Usage Restrictions:
CRITICAL: You must use models with ≤1B parameters.
Allowed models:
You must design a comprehensive benchmark to evaluate your model's code generation capabilities. Your benchmark should demonstrate thoughtful consideration of what makes good semiconductor simulation code.
Key Point: The quality of your benchmark design is as important as your model's performance. A well-designed benchmark demonstrates deep understanding of the problem domain and rigorous evaluation methodology.
| Component | Points | Description |
|---|---|---|
| Benchmark Design & Evaluation New | 30 | Quality and rigor of custom benchmark design, evaluation metrics, and results analysis |
| Implementation & Methodology | 30 | Training approach and technical implementation |
| Presentation | 20 | Live demonstration and explanation |
| Documentation & Code Quality | 20 | Technical report and code organization |
| Total | 100 |
Graduate students (CSC 575): Higher expectations for methodology sophistication, literature review, and analysis depth.
Submit via course website:
Deadline: December 12, 11:59 PM
Allowed:
Not Allowed: