DUAL AGGREGATION TRANSFORMER FOR IMAGE SUPER-RESOLUTION

BASED ON THE ABOVE IDEA, WE PROPOSE A NOVEL TRANSFORMER MODEL, DUAL AGGREGATION TRANSFORMER (DAT), FOR IMAGE SR. OUR DAT AGGREGATES FEATURES ACROSS SPATIAL AND CHANNEL DIMENSIONS, IN THE INTER-BLOCK AND INTRA-BLOCK DUAL MANNER.

FOODSAM: ANY FOOD SEGMENTATION

REMARKABLY, THIS PIONEERING FRAMEWORK STANDS AS THE FIRST-EVER WORK TO ACHIEVE INSTANCE, PANOPTIC, AND PROMPTABLE SEGMENTATION ON FOOD IMAGES.

LANGUAGE MODELS ARE FEW-SHOT LEARNERS

BY CONTRAST, HUMANS CAN GENERALLY PERFORM A NEW LANGUAGE TASK FROM ONLY A FEW EXAMPLES OR FROM SIMPLE INSTRUCTIONS - SOMETHING WHICH CURRENT NLP SYSTEMS STILL LARGELY STRUGGLE TO DO.

FINE-TUNING LANGUAGE MODELS FROM HUMAN PREFERENCES

MOST WORK ON REWARD LEARNING HAS USED SIMULATED ENVIRONMENTS, BUT COMPLEX INFORMATION ABOUT VALUES IS OFTEN EXPRESSED IN NATURAL LANGUAGE, AND WE BELIEVE REWARD LEARNING FOR LANGUAGE IS A KEY TO MAKING RL PRACTICAL AND SAFE FOR REAL-WORLD TASKS.