Glad to say that my newest research is available at ArXiv!

It might interest you if you're curious about using metalearning for small tabular datasets.
arxiv.org/abs/2304.03543

HyperTab: Hypernetwork Approach for Deep Learning on Small Tabular Datasets

Deep learning has achieved impressive performance in many domains, such as computer vision and natural language processing, but its advantage over classical shallow methods on tabular datasets remains questionable. It is especially challenging to surpass the performance of tree-like ensembles, such as XGBoost or Random Forests, on small-sized datasets (less than 1k samples). To tackle this challenge, we introduce HyperTab, a hypernetwork-based approach to solving small sample problems on tabular datasets. By combining the advantages of Random Forests and neural networks, HyperTab generates an ensemble of neural networks, where each target model is specialized to process a specific lower-dimensional view of the data. Since each view plays the role of data augmentation, we virtually increase the number of training samples while keeping the number of trainable parameters unchanged, which prevents model overfitting. We evaluated HyperTab on more than 40 tabular datasets of a varying number of samples and domains of origin, and compared its performance with shallow and deep learning models representing the current state-of-the-art. We show that HyperTab consistently outranks other methods on small data (with a statistically significant difference) and scores comparable to them on larger datasets. We make a python package with the code available to download at https://pypi.org/project/hypertab/

arxiv.org

One free message per day to GPT-4 and Claude+
---
RT @poe_platform
Today we are launching Poe subscriptions, which will provide paying users with access to bots based on two powerful new language models: GPT-4 from OpenAI and Claude+ from Anthropic.
twitter.com/poe_platform/statu

Gpt4 is cool and all, but remember that chatbots aren't limited to OpenAI
---
RT @AnthropicAI
After working for the past few moths with key partners like @NotionHQ, @Quora, and @DuckDuckGo, we’ve been able to carefully test out our systems in the wild. We are now opening up access to Claude, our AI assistant, to power businesses at scale.
twitter.com/AnthropicAI/status

I'd expect the much more accessible flan-t5 to make the waves that Llama is making instead. Is their difference in quality that big?

Show thread

Apparently @gpt_index is rebranding itself as LLamaIndex - but isn't Llama model hardly usable due to its licensing?

On the field of ChatGPT's competitors it's not nearly as impressive as @AnthropicAI's Claude, but it's a step in the right direction
---
RT @_philschmid
The first open-source ChatGPT alternative got released! 🚀@togethercompute released a 20B chat-GPT model on Apache-2.0 🗣🆕

You can try it for free on Hugging Face. 😍

Demo: huggingface.co/spaces/together
Model: huggingface.co/togethercompute
twitter.com/_philschmid/status

RT @_philschmid
The first open-source ChatGPT alternative got released! 🚀@togethercompute released a 20B chat-GPT model on Apache-2.0 🗣🆕

You can try it for free on Hugging Face. 😍

Demo: huggingface.co/spaces/together
Model: huggingface.co/togethercompute
Announcement: together.xyz/blog/openchatkit

Wszystkim laureatom przypominam o obowiązku używania hasztagów.

I've created an open source meeting summarizer and deployed it to Huggingface. It accepts texts of arbitrary length, going beyond the default token limit.

huggingface.co/spaces/wwydmans

Here are some notes taken from Hamlet:

RT @theworldshesaw
Imagine opening a @Nature article that has always been free and suddenly finding it paywalled. That’s what just happened to almost every Comment, News, and @NatureFutures article. This is disastrous. We NEED these freely available to the public (1/n)

Wygląda na to, że wśród pojawią się moje badania z XAI w biotechu.
W końcu wszyscy chcemy być zrozumiani, białka też na to zasługują

A healthy competition for ChatGPT is finally here! Claude by @AnthropicAI is available to use for everyone at poe.com, created by @Quora.

In a few years we're going to look at this problem the same way that we now look at SQL injections
---
RT @nearcyan
Indirect Prompt Injection: Turning Bing Chat into a Data Pirate

by modifying a website that bing chat reads alongside a user, the chat agent is able to have its goals modified by that site, unbeknownst to the user

demo: greshake.github.io/

arxiv: arxiv.org/abs/2302.12173
twitter.com/nearcyan/status/16

Nikt się tego nie spodziewał, ale "status: rozpatrywany" wskazuje na to, że Perły Nauki znowu mają opóźnienia względem opóźnionego terminu.

RT @CAMDA_conf
Try your best & take the challenge! First up: Track anti-microbial resistance in data & find the origin of resistant isolates from their ! Subscribe for more challenges 👉 camda.info @metasub @JimGolden23 @Niamh_Oh @mason_lab @GeoSeeq

I think I've found the one weird trick for 10xing my research career.

There is an unpleasant lack of metagenomic datasets on @huggingface so I've decided to add some, starting from Feng Q. et al.'s "Gut microbiome development along the colorectal adenoma-carcinoma sequence " publication.

huggingface.co/datasets/wwydma

RT @RadekKar
✈️ Krajowe loty poselskie to szara strefa wydawania publicznych pieniędzy.

Posłowie nie muszą raportować:

➡️ skąd lecą,
➡️ dokąd lecą,
➡️ i w jakim celu.

W tej kadencji na ten środek transportu wydaliśmy 12 720 486,36 zł.

👉 Zapraszam na materiał: youtu.be/MvMON3h8TsM

🙃
---
RT @LOkruszek
Ponieważ w sobotę minął rok od złożenia naszego wniosku w Perłach Nauki mieliśmy dziś małą imprezę rocznicową w labie razem z ówczesną magistrantką a dziś już doktorantką Moniką Malon! Dziękujemy @MEIN_GOV_PL za tą przygodę, czekamy na drugą rocznicę!
twitter.com/LOkruszek/status/1

Show older
Qoto Mastodon

QOTO: Question Others to Teach Ourselves
An inclusive, Academic Freedom, instance
All cultures welcome.
Hate speech and harassment strictly forbidden.