Study: Artificial Intelligence(AI)/AI: MultiModal
[논문리뷰] CLIP: Vision Language Multimodal dataset - CLIP: Learning transferable visual models from natural language supervision
💡 본 문서는 'CLIP: Learning transferable visual models from natural language supervision' 논문을 정리해놓은 글입니다.해당 논문은 CLIP 같은 멀티모달 모델의 language embedding을 NeRF 안에 집어넣어 NeRF를 Multi Modal로 확장 가능성을 보여준 논문이니 참고하시기 바랍니다. - Project: https://www.lerf.io/ - Paper: https://arxiv.org/abs/2303.09553 - Github: https://github.com/kerrj/lerf - Dataset: https://drive.google.com/drive/folders/1vh0mSl7v29yaGsxleadcj-LCZO..