SEPARATE ANYTHING YOU DESCRIBE

IN THIS WORK, WE INTRODUCE AUDIOSEP, A FOUNDATION MODEL FOR OPEN-DOMAIN AUDIO SOURCE SEPARATION WITH NATURAL LANGUAGE QUERIES.

PLATYPUS: QUICK, CHEAP, AND POWERFUL REFINEMENT OF LLMS

WE PRESENT A FAMILY OF FINE-TUNED AND MERGED LARGE LANGUAGE MODELS (LLMS) THAT ACHIEVES THE STRONGEST PERFORMANCE AND CURRENTLY STANDS AT FIRST PLACE IN HUGGINGFACE'S OPEN LLM LEADERBOARD AS OF THE RELEASE DATE OF THIS WORK.

CoDeF: CONTENT DEFORMATION FIELDS FOR TEMPORALLY CONSISTENT VIDEO PROCESSING

WE PRESENT THE CONTENT DEFORMATION FIELD F AS A NEW TYPE OF VIDEO REPRESENTATION, WHICH CONSISTS OF A CANONICAL CONTENT FIELD AGGREGATING THE STATIC CONTENTS IN THE ENTIRE VIDEO AND A TEMPORAL DEFORMATION FIELD RECORDING THE TRANSFORMATIONS FROM THE CANONICAL IMAGE (I. E., RENDERED FROM THE CANONICAL CONTENT FIELD) TO EACH INDIVIDUAL FRAME ALONG THE TIME AXIS. GIVEN A TARGET VIDEO, THESE TWO FIELDS ARE JOINTLY OPTIMIZED TO RECONSTRUCT IT THROUGH A CAREFULLY TAILORED RENDERING PIPELINE. WE ADVISEDLY INTRODUCE SOME REGULARIZATIONS INTO THE OPTIMIZATION PROCESS, URGING THE CANONICAL CONTENT FIELD TO INHERIT SEMANTICS (E. G., THE OBJECT SHAPE) FROM THE VIDEO. WITH SUCH A DESIGN, F NATURALLY SUPPORTS LIFTING IMAGE ALGORITHMS FOR VIDEO PROCESSING, IN THE SENSE THAT ONE CAN APPLY AN IMAGE ALGORITHM TO THE CANONICAL IMAGE AND EFFORTLESSLY PROPAGATE THE OUTCOMES TO THE ENTIRE VIDEO WITH THE AID OF THE TEMPORAL DEFORMATION FIELD. WE EXPERIMENTALLY SHOW THAT F IS ABLE TO LIFT IMAGE-TO-IMAGE TRANSLATION TO VIDEO-TO-VIDEO TRANSLATION AND LIFT KEYPOINT DETECTION TO KEYPOINT TRACKING WITHOUT ANY TRAINING. MORE IMPORTANTLY, THANKS TO OUR LIFTING STRATEGY THAT DEPLOYS THE ALGORITHMS ON ONLY ONE IMAGE, WE ACHIEVE SUPERIOR CROSS-FRAME CONSISTENCY IN PROCESSED VIDEOS COMPARED TO EXISTING VIDEO-TO-VIDEO TRANSLATION APPROACHES, AND EVEN MANAGE TO TRACK NON-RIGID OBJECTS LIKE WATER AND SMOG. PROJECT PAGE CAN BE FOUND AT IMAGE-TO-IMAGE TRANSLATION KEYPOINT DETECTION +1

OCTOPACK: INSTRUCTION TUNING LARGE LANGUAGE MODELS

WE BENCHMARK COMMITPACK AGAINST OTHER NATURAL AND SYNTHETIC INSTRUCTIONS (XP3X, SELF-INSTRUCT, OASST) ON THE 16B PARAMETER STARR MODEL, AND ACHIEVE STATE-OF-THE-ART PERFORMANCE AMONG MODELS NOT TRAINED ON OPENAI OUTPUTS, ON THE HUMANEVAL PYTHON BENCHMARK (46. 2% PASS@1).