Discussion about this post

User's avatar
mohamed sheded's avatar

Thanks this is Very helpful, how you face the challenge of rate limit for longer video frames

like 120 frame --> "gpt 4o mini" will exceed the limit rate, right ?

and where is the pixel table storing this data , is that persistent so i can get back to it ?

Expand full comment
Valentin Jimenez's avatar

Nice! You are digging deep into the papers. Awesome! I have a question, in pixeltable, we are using the model CLIP, particularly in the caption creation. However, in one of the articles, it's mentioned that BLIP has more powerful capabilities, right? BLIP is currently not available in pixeltable, the available models can be found here:

https://docs.pixeltable.com/docs/integrations/models#clip-models

To my understanding, BLIP is more powerful, at least in the sense that if we provide more noisy data, it still will produce good results, if I understood the article correctly. So, if we used BLIP, I guess we could reduce the video quality even more, in order to reduce the costs when calling the gpt-4o-mini model for captions. Is my understanding correct?

Expand full comment
9 more comments...

No posts