Hi ,
First of all thank you for hosting this competion it is an interesting challange .
I am focusing on Track B so all questions related to that
My assumption that Phase 1 score doesn't really count as this is aimed for understanding the enviroment.
Phase1 question:
1. Will the phase one gold truth be released prior to phase 2?
Phase 2 question :
2. I assume in Phase 2 the we cannot work with locally hosted enviroemtn and have to use the cloud server can you confirm ?
Phase 3 questions :
1, What it expected infernece speed in phase 3 (or what is the exact GPU will be used )?
2, Is memory across questions permissable ?
3, Is topology discovery prior questions allowed ?
4. Are there any resource limits — container CPU/RAM/GPU in Phase 3 grading?
5. LORA adapters expected in bf16 , or other quantization can be used as well ?
6. Can quantized lower precision version be used in phase 3 ( can a merged model be provided ) ?
Thank you very much in advance
L
Hi,
Phase 1 question: 1. Yes. We plan to release the golden ground truth of Phase 1 when Phase 1 is finished. Phase 2 question: 1. Yes. Phase 2 only allow api call from cloud service Phase 3 question: 1. We prefer to not share this information. Everyone will have access to the same resources. 2. No. Each question has an independent context. 3. Yes. We do not interfere with the way agents solve problems. 4. We prefer to not share this information. Everyone will have access to the same resources. 5. Any parameter precision is acceptable. 6. Any parameter precision is acceptable.
Hi Antonio, Thank you for the response !
Hi @AntonioDeDomenico, thanks for the earlier clarifications. A few follow-ups specifically on Phase 3 execution, to help us package submissions correctly:
I understand you may not want to disclose exact GPU model or VRAM. Even partial answers on the ecosystem (CUDA vs CANN) and the submission contract (weights-only vs full-container) are what we need most. Thanks!
1. CUDA 2. Please submit the full runnable environment with one-click button flow and make sure the environment can run successfully. The GPU deployment code and the Agent execution code need to be separated, because one is deployed on the GPU server and the other on the CPU server. 3. Please refer to question 2 4. Please do not worry about tool-call latency. The latency will be short enough not to affect the overall runtime.