Code Monkey home page Code Monkey logo

awesome-embodied-ai's Introduction

Awesome-Embodied-AI

Awesome Contrib

A curated list of awesome papers on Embodied AI and related research/industry-driven resources, inspired by awesome-computer-vision.

Embodied AI has led to a new breakthrough, and this repository will keep tracking and summarizing the research or industrial progress.

- Contribution is highly welcome and feel free to submit a pull request or contact me.

If you find this repository helpful, please consider Stars ⭐ or Sharing ⬆️.

Contents

  • Data Interpreter: An LLM Agent For Data Science
  • Communicative Agents for Software Development
  • Exposing Limitations of Language Model Agents in Sequential-Task Compositions on the Web
  • Experiential Co-Learning of Software-Developing Agents
  • EASYTOOL: Enhancing LLM-based Agents with Concise Tool Instruction
  • A survey of embodied ai: From simulators to research tasks
  • Embodied AI in education: A review on the body, environment, and mind
  • Agent ai: Surveying the horizons of multimodal interaction
  • Learning to Generate Context-Sensitive Backchannel Smiles for Embodied AI Agents with Applications in Mental Health Dialogues
  • Alexa arena: A user-centric interactive platform for embodied ai
  • Artificial intelligence education for young children: A case study of technology‐enhanced embodied learning
  • Embodiedgpt: Vision-language pre-training via embodied chain of thought
  • Multimodal embodied interactive agent for cafe scene
  • The Essential Role of Causality in Foundation World Models for Embodied AI
  • A Survey on Robotics with Foundation Models: toward Embodied AI
  • Where are we in the search for an artificial visual cortex for embodied intelligence?
  • A Comprehensive Survey of Cross-Domain Policy Transfer for Embodied Agents
  • The sense of agency in human–AI interactions
  • " Don't forget to put the milk back!" Dataset for Enabling Embodied Agents to Detect Anomalous Situations
  • Vision-Language Navigation with Embodied Intelligence: A Survey
  • Hierarchical Auto-Organizing System for Open-Ended Multi-Agent Navigation
  • Velma: Verbalization embodiment of llm agents for vision and language navigation in street view
  • Spatially-Aware Transformer Memory for Embodied Agents
  • VELMA: Verbalization Embodiment of LLM Agents for Vision and Language Navigation in Street View
  • Embodied Human Activity Recognition
  • LoTa-Bench: Benchmarking Language-oriented Task Planners for Embodied Agents
  • EDGI: Equivariant diffusion for planning with embodied agents
  • Large Multimodal Agents: A Survey
  • Egocentric Planning for Scalable Embodied Task Achievement
  • EnvGen: Generating and Adapting Environments via LLMs for Training Embodied Agents
  • Human-agent teams in VR and the effects on trust calibration
  • Talk with Ted: an embodied conversational agent for caregivers
  • MOPA: Modular Object Navigation With PointGoal Agents
  • Embodied Conversational Agents for Chronic Diseases: Scoping Review
  • Towards anatomy education with generative AI-based virtual assistants in immersive virtual reality environments
  • Improving Knowledge Extraction from LLMs for Task Learning through Agent Analysis
  • A Survey on Large Language Model-Based Game Agents
  • Autort: Embodied foundation models for large scale orchestration of robotic agents
  • Towards Heterogeneous Multi-Agent Systems in Space
  • Embodied Machine Learning
  • Penetrative ai: Making llms comprehend the physical world
  • WebVLN: Vision-and-Language Navigation on Websites
  • Generating meaning: active inference and the scope and limits of passive AI
  • RoboHive: A Unified Framework for Robot Learning
  • Grounded Decoding: Guiding Text Generation with Grounded Models for Embodied Agents
  • Turing Test in the Era of LLM
  • Generative Models for Decision Making
  • AgentScope: A Flexible yet Robust Multi-Agent Platform:octocat:
  • MMMU: A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for Expert AGI:octocat:
  • MuLan: Multimodal-LLM Agent for Progressive Multi-Object Diffusion:octocat:
  • Vision-Language Navigation with Embodied Intelligence: A Survey
  • An Interactive Agent Foundation Model
  • UFO:A UI-Focused Agent for Windows OS Interaction:octocat:

Survey

2023

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.