Show newer

Locke, Berkeley, and Hume, emphasized the idea that experience is the ultimate source and justification of knowledge.

Critics argue that Popper's criterion is both too restrictive (excluding some legitimate scientific claims) and too permissive (allowing some pseudoscientific theories)

Popper emphasized the asymmetry between confirmation and falsification, arguing that while no finite amount of evidence can conclusively prove a theory, a single counterexample can potentially disprove it.

A critical step for RAG with large documents is chunking – breaking text into smaller, manageable parts. This ensures relevant sections can be retrieved and fit within the LLM's context window

Building a chat application over your own data often involves RAG and a vector database. Vector databases store and allow efficient searching of data embeddings.

Retrieval-Augmented Generation (RAG) is a powerful technique to ground LLMs on external data. This enhances the relevance of responses and helps reduce hallucinations by providing context.

Control the randomness of LLM outputs using temperature or top_p parameters. Lower values yield deterministic results, while higher values increase creativity and diversity.

Large Language Models (LLMs) process text by converting it into tokens. For English, a token is roughly 4 characters or 0.75 words. Managing tokens is vital for cost and performance.

Isaac Asimov introduced his famous Three Laws of Robotics in 1942, establishing a hierarchy for robotic decision-making, like prioritizing preventing harm over following orders.

Imagine assessment that focused on:: habits, comparison to others, & ability to create valued work.

There's growing recognition that diverse assessment data is needed for a complete picture of student learning, beyond traditional tests.

Today, I encountered what appears to be Chinese characters in a generative AI response.

Before AI can be widely adopted, people must trust it, especially that it can make accurate and fair decisions. AI should be aware of and aligned with human values.

Biased humans contributing to AI data is another source of bias. Human discrimination in areas like the labor market can become manifest in computer algorithm.

Facial recognition algorithms developed in East Asia performed better on Asian subjects, while Western algorithms performed better on White subjects. This discrepancy is attributed to different racial distribution in training sets.

An AI might correlates a zip code with better employee performance and incorrectly assumes the zip code causes the performance.

One way AI decisions become biased is by confusing correlation with causation. Just because two variables change together doesn't mean one causes the other.

AI is increasingly used for critical decisions in hiring, loans, medicine, and more. While AI can perform certain tasks more accurately than humans, using computers doesn't automatically eliminate bias.

Negotiating technology sufficiency in schools involves balancing the capacity of systems/devices, the number of devices, how they are available, and teacher preparation.

Learning outside school is often socially embedded, interest-driven, and opportunity-oriented. Students arriving in classrooms are active & independent learners due to digital experiences.

Show older
Qoto Mastodon

QOTO: Question Others to Teach Ourselves
An inclusive, Academic Freedom, instance
All cultures welcome.
Hate speech and harassment strictly forbidden.