This Politico article argues that #NMT (Neural Machine Translation) was one of the major drivers of European unity against the Russian invasion of Ukraine.
Not sure I fully buy the argument, but MT is probably one of the best examples of #NLProc / #AI that benefits society / #AIforGood.
https://www.politico.com/news/magazine/2023/02/03/europe-putin-ukraine-google-translate-00079301
h/t @daanvanesch
CALL FOR PAPERS: Research and Innovation Track
We welcome papers on novel scientific research and/or innovations relevant to #SemanticWeb, #KnowledgeGraphs, #AI, #ML, #NLP and more
Deadlines:
🗓️Abstracts: May 09
🗓️Papers: May 16
For more info: 🌐https://2023-eu.semantics.cc/page/cfp_rev_rep
Machbarkeitsstudie: KI-Leuchtturmprojekt in Deutschland möglich
ChatGPT kommt aus den USA. KI-Experten in Deutschland sehen darin ein Problem und fordern einen Kraftakt zur Wahrung der digitalen Souveränität.
#Bundeswirtschaftsministerium #ChatGPT #Deutschland #Europa #KünstlicheIntelligenz #LEAM #Sprachmodelle #Supercomputer #USA #digitaleAssistenten
In eigener Sache: heise online zieht auf eigene Mastodon-Instanz
Das Chaos bei Twitter hält an und die Mastodon profitiert weiter. Heise Medien betreibt in dem Fediverse-Netzwerk nun eine eigene Instanz.
#Fediverse #Heise #Mastodon #SocialMedia #Twitter #TwitterÜbernahme #heiseonline
Do you love #selfhosting? What about providing service to the public via #Codeberg?
We are looking for maintainers that take on adding code search features to our #Forgejo instance to reduce the load on the existing infrastructure team and bring this project forward.
Please see https://codeberg.org/Codeberg/Community/issues/904 if you are interested.
We are looking forward to your contributions. Thank you a lot!
The most interesting thing about #ChatGPT that no one is talking about is how the future will be systems talking to each other with imprecise protocols but they’re still able to understand
And the year has barely started!
RT @MishaLaskin@twitter.com
In-context RL at scale. After online pre-training, the agent solves new tasks entirely in-context like an LLM and works in a complex domain. One of the most interesting RL results of the year. https://twitter.com/FeryalMP/status/1616035293064462338
🐦🔗: https://twitter.com/MishaLaskin/status/1616066421582176258
Recent events have demonstrated how crucial resilience (e.g. of supply chains) is for our society. Semantic technologies can play a crucial here.
We will organize the #D2R2 (Linked Data-driven Resilience Research) #workshop at @eswc_conf@twitter.com
in May 2023 in Crete. We are looking forward to your contribution. Submission deadline is March 9. Check more details on our event page! https://d2r2.aksw.org #ESWC23 #CoyPu_Project #Resilience #LinkedData #cfp
The artificial-intelligence chatbot ChatGPT can write fake abstracts that scientists have trouble distinguishing from those written by humans. Increasing sophistication of chatbots could undermine research integrity and accuracy, researchers fear.
https://www.nature.com/articles/d41586-023-00056-7
#chatbots #ChatGPT #research #AI #ArtificialIntelligence
via @Nature
Hello NLP researchers around the globe! All ACL major conferences (@aclmeeting, @eaclmeeting, @aaclmeeting, and @emnlpmeeting) now have an account here. Please spread it word! #NLPRoc
I found the papers "Scaling Laws for Neural Language Models" (OpenAI, 2020) and "Training Compute-Optimal Large Language Models" (DeepMind, 2022) interesting:
https://arxiv.org/pdf/2001.08361.pdf
https://arxiv.org/pdf/2203.15556.pdf
They do a LOT of experiments training large language models (causal transformers) with varying hyperparameters, in particular model size, shape, batch size, and training data set size over many orders of magnitude. 1/?
DeepMind's paper refutes this last claim, and finds that both are equally useful.
The differences between DeepMind & OpenAI's papers matter in terms of forecasting how big LLMs need to get. They arrived at these different conclusions because DeepMind did more learning rate tuning. This blog post https://severelytheoretical.wordpress.com/2022/07/18/thoughts-on-the-new-scaling-laws-for-large-language-models/ hypothesizes that DeepMind's paper might also be not doing enough hyperparameter tuning, and the scaling law may be less severe, perhaps not even a power law.
3/3
On #TheDataExchangePod I speak with Mark Chen, Research Scientist at OpenAI. We discuss the evolution of DALL·E, key research developments that led to DALL·E 2, data sources, safety measures, ML models needed for its success. #machinelearning #dalle2 #dalle #AI #generativeai https://thedataexchange.media/exploring-dalle-2/
I do however have high hopes for #blogic and RDF+Surfaces to make the interpretation of RDF vocabularies interoperable across organizations
Very interesting essay on LLMs, their limitations, and their future by @yoavgo!
https://gist.github.com/yoavg/59d174608e92e845c8994ac2e234c8a9
The latest issue of 'Ahead of AI' is now available!
This edition covers my top 10 papers of the year, as well as trends in the AI industry, notable developments in open source projects, and my personal yearly review routine.
Check it out at the link below and have a happy new year!
https://magazine.sebastianraschka.com/p/ahead-of-ai-4-a-big-year-for-ai
How good of a BERT can one get in ONE DAY on ONE GPU?
With all the recent studies about scaling compute up, this paper takes a refreshing turn and does a deep dive into scaling down compute.
It's well written, stock full of insights. Here is my summary and my opinions.
https://arxiv.org/abs/2212.14034
🧶 1/N
With the advent of #ChatGPT, everyone is talking about large language models. But how do they work? Initially, such models were trained to complete sentences.
But they exhibit exciting capabilities that can be invoked by feeding them "prompts."
Read our Prompt Engineering Guide for a quick overview of the current state of this field.
#nlproc #gpt #llm
https://www.inovex.de/de/blog/prompt-engineering-guide/
Scikit-learn 1.2 is out: https://github.com/scikit-learn/scikit-learn/releases/tag/1.2.0
Was an eventful December & I totally missed the new release of my favorite #machinelearning library!
My personal highlights are around the HistGradientBoostingClassifier (if you haven't used it yet, it's a LightGBM impl that works really well)
It now supports
1. interaction constraints (in trees, features that appear along a particular path are considered as "interacting")
2. class weights
3. feature names for categorical features