Would you like to see your presentation here, made available to a global audience of researchers?
Add your own presentation or have us affordably record your next conference.
Large language models (LLMs) concentrate substantial knowledge in specialized domains due to extensive pretraining and instruction tuning, and they are now central to commercial and scientific practice. Yet access is usually limited to costly, rate limited interfaces, which motivates methods that can extract targeted domain knowledge with minimal querying effort. A further challenge is that the target domain may be unknown in advance, so naive or generic prompts waste queries and fail to expose the underlying concepts and relations that structure the domain. In this work, we introduce a query efficient approach for domain specific knowledge stealing from black box language models. Rather than issuing random questions or generic templates, our framework performs self directed exploration that lets the model find the direction and mine domain knowledge by itself. Starting from a small and diverse seed, it discovers salient domain entities and induces their relations through structured question families that elicit definitional, functional, and compositional information. A feedback driven controller analyzes the errors and uncertainty of the extracted student model and uses this signal to refine subsequent queries, all without any prior domain knowledge or external resources. We evaluate the method in two expert centric settings, medicine and finance, and observe consistently better performance while requiring significantly fewer queries.