{"id":970518,"date":"2023-09-27T20:33:57","date_gmt":"2023-09-28T03:33:57","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-academic-program&p=970518"},"modified":"2025-02-21T04:56:10","modified_gmt":"2025-02-21T12:56:10","slug":"microsoft-research-asia-startrack-program","status":"publish","type":"msr-academic-program","link":"https:\/\/www.microsoft.com\/en-us\/research\/academic-program\/microsoft-research-asia-startrack-program\/","title":{"rendered":"Microsoft Research Asia StarTrack Scholars"},"content":{"rendered":"\n\n
<\/p>\n\n\n\n\n\n\n
Microsoft Research Asia (MSRA) StarTrack Scholars Program is a visiting researcher program dedicated to empowering young scholars from across the globe. The program extends an exclusive invitation to outstanding young faculty members worldwide, offering them a three-month research visit opportunity at Microsoft Research Asia. The primary objective of MSRA StarTrack Scholars Program is to foster close academic exchange and collaboration between Microsoft Research Asia and young scholars from esteemed international universities and academic research institutions. Leveraging a world-class international industrial research platform, the program aims to empower young scholars to explore the new paradigm of computing for the coming decades, embrace interdisciplinary and cross-domain research, create breakthrough technologies with significant impact, and address major technical, industrial, and societal challenges.<\/p>\n\n\n\n
Join Microsoft Research Asia\u2019s StarTrack Scholars Program and embark on a journey of brilliance and possibilities:<\/p>\n\n\n\n
If you have any questions, please email Ms. Yanxuan Wu, program manager of the Microsoft Research Asia StarTrack Scholars, at\u202fv-yanxuanwu@microsoft.com<\/a>. <\/em><\/p>\n\n\n\n\n\n With the rise of large-scale AI models, such as Large Language Models (LLMs), we are witnessing a transformation in how these technologies are integrated into various aspects of our society. These models stand out for two key reasons: 1) General-purpose functionality: LLMs can perform a wide range of tasks, from translation and question answering to code completion and more; 2) Human-like competence: They have demonstrated the ability to perform many tasks at a level comparable to human, making them accessible and versatile tools for various domains.<\/p>\n\n\n\n While these powerful models offer significant societal benefits, they also introduce unforeseen challenges. These challenges arise not only from technical complexities but also from the broader social implications of widespread AI adoption. As Brad Smith aptly noted, \u201cThe more powerful the tool, the greater the benefit or damage it can cause.\u201d<\/p>\n\n\n\n To ensure that AI\u2019s integration into society is harmonious, synergistic, and resilient\u2014minimizing any potential side effects\u2014it is critical to foster Societal AI research. This emerging field prioritizes a multi-disciplinary approach, bringing together computer science and social science to address the complex dynamics of AI\u2019s role in shaping our world.<\/p>\n\n\n\n We invite researchers from both fields to join us in this exciting endeavor. Together, we can explore innovative solutions that ensure the responsible and equitable advancement of AI technologies.<\/p>\n\n\n\n\n\n Xing Xie<\/a> (Engaging Lead) We warmly invite researchers from both computer science and social sciences to join us in exploring the exciting frontier of societal AI. As we stand at the intersection of technological innovation and social impact, this collaboration provides a unique opportunity to bridge diverse fields of expertise. We are confident that this interdisciplinary approach will not only advance our knowledge of intelligence but also help shape AI technologies that are aligned with human values and needs. Ultimately, this partnership holds great promise for generating long-term benefits for human society, ensuring that AI serves as a force for positive social change.<\/p>\n<\/div>\n<\/div>\n\n\n\n Fangzhao Wu<\/a> We will work together with the visiting scholars on two research projects. The first one is improving the safety, privacy, and reliability of LLMs, and the second one is studying the impact of LLMs on humans and society, as well as how to maximize the positive impact and minimize the negative impact.<\/p>\n<\/div>\n<\/div>\n\n\n\n Xiaoyuan Yi<\/a> In this collaboration, we aim to integrate insights from social sciences, psychology, and related fields, adopting an interdisciplinary perspective to envision the ideal form of the next-generation AI. Our goal is to develop methods for evaluating, analyzing, and understanding the intrinsic driving factors of big models\u2014such as their values. Additionally, we seek to predict and align the behavior of these models based on their traits like values, maximizing the societal well-being that AI technology can offer in a future where humans and machines coexist symbiotically.<\/p>\n<\/div>\n<\/div>\n\n\n\n Jianxun Lian<\/a> Through this collaboration, we aim to explore the synergy between AI and social science. For instance, how can we leverage social science to evaluate, interpret, and shape AI behaviors? Conversely, how can a well-aligned AI transform the field of social science research? To answer these questions, we need to gain a deeper understanding of AI behaviors from an interdisciplinary perspective, particularly on those traits that reflect human-likeness, such as emotions, motives, preferences, beliefs, social intelligence, and theory of mind.<\/p>\n<\/div>\n<\/div>\n\n\n\n We envision a future where multimedia and advanced AI are seamlessly integrated into our daily life, building a hybrid world, and significantly enhancing humans\u2019 capabilities and experiences. Media Foundation is a cutting-edge research theme that aims to investigate the foundations of multimedia and advanced AI from multiple disciplines while ensuring responsible AI practices. Humans are amazing learners who can acquire a wide range of skills and knowledge from multiple modalities, such as vision, hearing, touch, and language. Advanced AI, like humans, also needs to learn from the real world. But how can we bridge the gap between the complex and noisy environment and the abstract and semantic representations of advanced AI? Media Foundation can transform the physical and virtual worlds into media tokens, which can be manipulated and communicated, thereby enhancing the learning and creativity of advanced AI and enabling novel multimodal applications. This research theme aims to explore the synergy between multimedia and advanced AI, covering topics such as neural codec, immersive telepresence, multimedia understanding, AIGC for user interfaces and HCI, 3D and computer vision, audio and speech, and deep learning fundamentals for multimedia.<\/p>\n\n\n\n For more detailed information, please refer to the article Media Founda<\/a>tion<\/a><\/em>.<\/p>\n\n\n\n\n\n Yan Lu<\/a> (Engaging Lead) We welcome applicants for media foundation research, which aims to build an intelligent environment through world model for benefiting humanity in the future. We seek candidates with a strong background in AI and multimedia, or related fields, including individuals with experience in AI foundaitons, neural codecs, computer vision, audio and speech, HCI, and other related domains. We also expect participants are passionate about exploring the paradigm shifts in research and are eager to contribute to the development of advanced AI systems that can enhance human learning and creativity.<\/p>\n<\/div>\n<\/div>\n\n\n\n Xiulian Peng<\/a> We welcome applications from individuals with expertise in audio\/speech, multi-modal learning, AI, and related disciplines, especially audio\/speech technologies. We anticipate that candidates will have a profound interest in exploring transformative developments within the realm of audio\/speech and multimedia research and be enthusiastic about contributing to the advancement of cutting-edge AI systems to elevate humans\u2019 capabilities and quality of life.<\/p>\n<\/div>\n<\/div>\n\n\n\n Shujie Liu<\/a> We welcome researchers with expertise in Artificial Intelligence Generated Content (AIGC), Multi-modal Large Language Models, Speech and Video Generation, and Human-Computer Interaction (HCI) to join our cutting-edge research programs. Our team is committed to pushing the boundaries of AI in creative content generation, enhancing how machines understand and produce human-like interactions across various modalities. We are particularly interested in scholars who seek to explore advanced AI techniques for multimodal fusion, adaptive interfaces, and human-centric AI applications. By fostering interdisciplinary collaborations, we aim to create groundbreaking technologies that redefine how humans interact with intelligent systems.<\/p>\n<\/div>\n<\/div>\n\n\n\n Bin Li<\/a> We welcome candidates with expertise in multimedia, AI, and related fields, such as neural codecs, computer vision, and deep learning. We value individuals passionate about investigating video research\u2019s paradigm shifts and contributing to advanced AI systems development, fostering human learning and creativity enhancement. We expect applicants to have a deep understanding of the field and are committed to pushing the boundaries of AI and multimedia research.<\/p>\n<\/div>\n<\/div>\n\n\n\n With the rapid advancements in AI and robotics, the development of highly intelligent robots capable of seamlessly interacting with the physical environment is becoming increasingly achievable. As the next AI\/AGI wave, embodied AI innovations promise to revolutionize various industries and significantly impact human life.<\/p>\n\n\n\n Although promising progress has been made, generalist robots and embodied AGI are still in their infancy. Embodied AI foundation models differ significantly from existing LLMs and VLMs. The ability to produce dense robot actions precisely and efficiently for physical interaction with general objects in the 3D spatial world is far beyond the capability of existing AI foundation models.<\/p>\n\n\n\n Our research aims to build a new generation of foundational embodied AI models with enhanced 3D spatial and physical proficiencies in perception, reasoning, and action. We explore vision-language-action model architectures, Internet-scale action data construction approaches, multimodal-sensory intelligence, and fundamental 3D computer vision techniques, among others. We are dedicated to making technical breakthroughs and creating new business opportunities for the company.<\/p>\n\n\n\n For more detailed information, please refer to the article Redefining robot intelligence<\/a><\/em>.<\/p>\n\n\n\n\n\n Baining Guo<\/a> (Engaging Lead) We invite researchers with a deep enthusiasm for Embodied AI to become part of our team. Embodied AI represents the next breakthrough in developing intelligent systems capable of perceiving, understanding, and interacting with the world in a more human-like way. If AI can talk and see, why not give it a body? If you have a strong research foundation in areas like foundation models, 3D computer vision, and robotics, we encourage you to join us. Together, let’s advance this new frontier in AI.<\/p>\n<\/div>\n<\/div>\n\n\n\n Jiaolong Yang<\/a> (Engaging Lead) I\u2019m eager to welcome professionals with a solid background in 3D computer vision, robotics, language models, reinforcement learning, video generation, and related fields. I expect our visiting researcher to have a strong passion in Embodied AI and be prepared to collaborate closely with our team during their visit. I also hope to build long-term partnerships with our visiting scholars, working together to produce influential contributions in the forthcoming surge of advancements in Embodied AI.<\/p>\n<\/div>\n<\/div>\n\n\n\n Lily Sun<\/a> (Engaging Lead) Embodied AI has recently emerged as one of the most promising fields, with immense potential for disruptive breakthroughs that could transform various industries and daily life. While significant progress has been made, many challenges remain in developing general foundation models for embodied AI that can empower diverse robots to solve problems across industrial and everyday scenarios. We warmly invite StarTrack scholars from around the world to join us in advancing the frontiers of this dynamic field and achieving impactful real-world outcomes!<\/p>\n<\/div>\n<\/div>\n\n\n\n Yaobo Liang<\/a> We welcome researchers with a strong passion for Embodied AI to join our team. Embodied AI is the next frontier in creating intelligent systems that can perceive, understand, and interact with the real world in a more human-like manner. When AI can talk and see, let’s give it a body! If you have a solid research background in foundation models,3Dcomputer vision, and robotics, we invite you to join us. Let’s work together to push this new AI frontier.<\/p>\n<\/div>\n<\/div>\n\n\n\n Yu Deng (opens in new tab)<\/span><\/a> I look forward to collaborating closely with visiting researchers on topics related to foundation models in Embodied AI, as well as 3D\/4D reconstruction and generation. I hope participants will have a solid background in 3D computer vision and graphics, along with a passion for tackling cutting-edge challenges in these fields. Let’s work together to advance research frontiers and make a big impact.<\/p>\n<\/div>\n<\/div>\n\n\n\n Fangyun Wei<\/a> I am eager to collaborate closely with visiting researchers on advancing embodied AI, with a particular focus on machine perception, reinforcement learning, and dexterous robotic grasping. Additionally, I look forward to exploring cutting-edge developments in vision-language-action foundation models, teleoperation, neural simulation environments, and the integration of intelligent robotic agents.<\/p>\n<\/div>\n<\/div>\n\n\n\n With the increasingly critical role of AI, the very foundation of computing is being redefined around AI. This transformation demands innovations across multiple dimensions of AI and computing systems. This research theme aims to call for research proposals addressing the fundamental challenges in computer systems in the era of AI. The topics include but are not limited to: (1) disruptive methods to significantly lower the bar of developing secure and reliable systems; (2) next generation software\/hardware architecture for future AI; (3) next generation storage\/database systems for AI; (4) Development and debugging tools for next generation intelligent systems; and (5) approaches for continual learning that enable systems to retain and adapt knowledge over time without extensive retraining. We particularly welcome unconventional, multi-discipline approaches to tackling the above challenges.<\/p>\n\n\n\n For more detailed information, please refer to the article Next-generation systems in the AI era<\/a><\/em>.<\/p>\n\n\n\n\n\n Fan Yang<\/a> (Engaging Lead) In the past decade, systems have played a critical role in fostering the unforeseen development of AI. We welcome young systems researchers to join us exploring the fundamental system principles in the era of AI.<\/p>\n<\/div>\n<\/div>\n\n\n\n Mike Chieh-Jan Liang<\/a> We are interested in using LLM-agents to continually evolve system design. System design goes beyond code completion, as it requires LLM-agents to (1)<\/em><\/strong> generate new ideas and (2)<\/em><\/strong> reason+refine potential ideas. Our current prototype mimics how right-brain and left-brain work. In addition to designing traditional systems, I also welcome collaborations to tackle AI infra design problems.<\/p>\n<\/div>\n<\/div>\n\n\n\n Ting Cao<\/a> We are at a pivotal moment where the LLM-centered economy is rapidly emerging, calling for a rethinking of hardware, systems, and algorithms to bring this vision into reality. Our team is focused on advancing these technologies to overcome the critical bottlenecks of LLM serving, such as the scaling of computing power, memory capacity and model capability\u2014especially for ubiquitous user devices. We invite researchers to join us in pushing the boundaries of what\u2019s possible.<\/p>\n<\/div>\n<\/div>\n\n\n\n Xian Zhang<\/a> We are working on neural-symbolic math reasoning. We welcome researchers passionate about boosting AI math reasoning through any potential approaches towards expert level. For example, expertise in the areas of mathematics with ties to AI and machine learning, including computer-assisted mathematics such as the use of proof assistants (e.g., Lean, Isabelle, Coq); specializing in one or more areas of high-performance computing, ML systems, programming languages, and formal methods.<\/p>\n<\/div>\n<\/div>\n\n\n\n Li Lyna Zhang<\/a> We are working on self-improving AI. We welcome researchers working on: (1) Optimizing LLM inference scaling laws; (2) Improving LLM reasoning capabilities, particularly in domains like math and code reasoning; (3) LLM post-training, such as supervised-finetuning, reinforcement learning from human feedback, or reward model training; (4) optimizing LLM long-context capabilities. Identify and address cutting-edge research challenges in the real applications of long-context LLMs.<\/p>\n<\/div>\n<\/div>\n\n\n\n Lingxiao Ma<\/a> We welcome researchers working on deep learning compilers to join us to explore the compiler techniques for future AI models and hardware<\/p>\n<\/div>\n<\/div>\n\n\n\n Shijie Cao<\/a> Our research focuses on addressing the critical efficiency challenges that large language models (LLMs) face as they continue to transform industries and everyday life. To truly make LLMs accessible on every device, innovative approaches in model-system-hardware co-design are essential. We explore cutting-edge techniques, such as low-bit precision and sparsity, and novel system optimizations and custom hardware designs specifically tailored for these techniques. We are excited to invite talented researchers to collaborate with us in pioneering the solutions that will shape the future of AI at scale.<\/p>\n<\/div>\n<\/div>\n\n\n\n Shuai Lu<\/a> We are working on LLM for code verification. We welcome faculties with interests and experiences in: (1) Using LLMs for code like code generation, automatic debugging, and code verification. Applying LLMs to system and software engineering problems. (2) Applying LLMs to reasoning tasks, designing and training models to solve domain-specific problems. (3) Formal verification and familiar with formal languages like Verus, Dafny, Isabelle, or Lean.<\/p>\n<\/div>\n<\/div>\n\n\n\n Baotong Lu<\/a> We believe in the critical role of the \u201cstorage component\u201d in the era of AI. Specifically, our research is dedicated to enhancing the efficiency of large language models (LLMs) in critical scenarios, such as long context inference, through hardware-algorithm co-design. We strive to understand and leverage the fundamental working principles of the large languages models to advance the model performance. We welcome researchers working on modern storage systems, e.g., database systems in disaggregated hardware, vector stores, storage for LLMs, to join us in the exploration.<\/p>\n<\/div>\n<\/div>\n\n\n\n Ran Shu<\/a> We are working on simulation of the large-scale AI systems. Specifically, we are focusing on cutting-edge simulators with extreme efficiency and fidelity. We welcome professors to collaborate on: (1) hardware \/ software co-design for network simulation acceleration; (2) efficient and accurate GPU \/ system bus \/ NIC simulation and modeling; (3) efficient and accurate traffic generation and monitoring in simulation.<\/p>\n<\/div>\n<\/div>\n\n\n\n Zhirong Wu<\/a> Training large language models is highly computationally intensive, and with the rapid expansion of global knowledge each year, updating LLMs incrementally without extensive retraining has become a pressing challenge. We invite applicants with interests in continual learning, mitigating catastrophic forgetting, neural plasticity, and efficient model adaptation to join us in tackling this timely issue.<\/p>\n<\/div>\n<\/div>\n<\/div>\n\n\n\n\n\n Our brain is one of the most complicated objects on this planet. Although scientists have spent hundreds of years trying to unveil our brain, there are still a lot of questions remaining unresolved. Understanding how the brain works is very important because it can help us better treat neurological and psychiatric disorders and develop more advanced artificial intelligence. In this theme, we aim to explore how to use AI to better understand the brain and apply the insights to improve brain health and design brain-inspired AI.<\/p>\n\n\n\n For more detailed information, please refer to the article Towards a synergy between AI and the brain<\/a><\/em>.<\/p>\n\n\n\n\n\n Dongsheng Li<\/a> (Engaging Lead) As AI is transforming the world, it is equally crucial to deepen our understanding of biological intelligence. The integration of AI and brain science research holds immense potential to drive innovation, enhance human well-being, and expand our knowledge of both artificial and biological intelligence. I am eager to collaborate with researchers in this field to revolutionize existing paradigms through interdisciplinary studies, advancing both AI and brain science in groundbreaking ways.<\/p>\n<\/div>\n<\/div>\n\n\n\n Yansen Wang<\/a> Our brain, encapsulated within a structure no larger than our two fists, is a marvel of complexity and power. It governs every facet of our thoughts, emotions, and actions, enabling feats from artistic expression to scientific innovation. Yet, much of its inner workings remain shrouded in mystery, with countless questions about consciousness, memory, and cognition still unanswered. Understanding the brain is not only essential for unraveling these mysteries but also for constructing interfaces between the brain and artificial intelligence. I hope to see a better synergy in the future. <\/p>\n<\/div>\n<\/div>\n\n\n\n Dongqi Han<\/a> A single neuron, while capable of carrying some information, cannot generate intelligence on its own. However, when many neurons connect and interact, they form complex networks\u2014whether in the biological brain or artificial neural networks\u2014capable of processing information, learning, and making decisions. The secret behind this collective intelligence lies in the emergent properties of interconnected systems, where the intricate patterns of connections and dynamic interactions between neurons give rise to cognition, perception, and learning. Unraveling the origin of intelligence requires research into these interactions, exploring how simple components combine to produce profound capabilities. <\/p>\n<\/div>\n<\/div>\n\n\n\n\n\n Large language models, led by GPT and followed by numerous other models, have demonstrated their strong capabilities in many areas from language-based tasks to reasoning and planning. LLMs have fueled the new wave of AI revolution that will generate deep impact on every aspect of the society. However, behind the LLMs, its basic mechanism can be simply explained as next word prediction — predicting the next word to be generated as accurate as possible after trained on a large amount of text data. What is the intrinsic connection between the seemly simple next word prediction mechanism and the amazingly intelligent capabilities on language processing, theory of mind, reasoning, planning, and so on? Is there any fundamental limitation of LLMs based on next word prediction? Currently there is still lack of principled understanding of these crucial questions.<\/p>\n\n\n\n In this proposal, we would like to solicit research collaborations on the principled and theoretical understanding of the capabilities of LLMs. The potential collaborations may cover a wide range of research topics, such as understanding the reasoning and planning capabilities of LLMs, theoretical explanation of the emergent behavior of LLMs, LLMs as compressors, enhancing LLMs with data-efficient training, new architectures, or paradigms for language models, etc. There is no established theory yet for large language models and generative AI systems in general, which means that the area is wide open and there are many opportunities to explore. The exploration may like to require interdisciplinary approaches and the integration with empirical-based approaches to be successful. We believe that developing a solid theoretical foundation of LLMs is imperative for the future development of safe, secure, and interpretable AI systems.<\/p>\n\n\n\n For more detailed information, please refer to the article Theoretical foundation of large language models<\/a><\/em>.<\/p>\n\n\n\n\n\n1. Societal AI<\/h5>\n\n\n\n
\n
<\/figure>\n\n\n\n
PARTNER RESEARCH MANAGER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
PRINCIPAL RESEARCHER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
SENIOR RESEARCHER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
SENIOR RESEARCHER<\/p>\n<\/div>\n\n\n\n2. Media Foundation<\/h5>\n\n\n\n
\n
<\/figure>\n\n\n\n
PARTNER RESEARCH MANAGER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
PRINCIPAL RESEARCH MANAGER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
PRINCIPAL RESEARCHER
<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
PRINCIPAL RESEARCHER<\/p>\n<\/div>\n\n\n\n3. Embodied AI and Large Action Models<\/h5>\n\n\n\n
\n
<\/figure>\n\n\n\n
DISTINGUISHED SCIENTIST<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
Principal Research Manager<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
DIRECTOR of MSR Accelerator\u2013Region & China<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
SENIOR RESEARCHER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
SENIOR RESEARCHER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
SENIOR RESEARCHER <\/p>\n<\/div>\n\n\n\n4. Next generation systems in the AI era<\/h5>\n\n\n\n
\n
<\/figure>\n\n\n\n
SR PRINCIPAL
RESEARCH MANAGER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
PRINCIPAL RESEARCHER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
PRINCIPAL RESEARCH MANAGER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
SENIOR RESEARCHER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
SENIOR RESEARCHER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
SENIOR RESEARCHER <\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
SENIOR RESEARCHER <\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
RESEARCHER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
RESEARCHER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
SENIOR RESEARCHER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
SENIOR RESEARCHER<\/p>\n<\/div>\n\n\n\n5. Towards a synergy between AI and brain<\/h5>\n\n\n\n
\n
<\/figure>\n\n\n\n
PRINCIPAL RESEARCH MANAGER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
SENIOR RESEARCHER<\/p>\n<\/div>\n\n\n\n<\/figure>\n\n\n\n
SENIOR RESEARCHER<\/p>\n<\/div>\n\n\n\n6. Theoretical Foundation of Large Language Models<\/h5>\n\n\n\n
\n