Contents^
Items^
Tell HN: The next generation of videogames will be great with midjourney
carson-katri/dream-textures (Stable Diffusion, Blender) https://github.com/carson-katri/dream-textures
nv-tlabs/GET3D https://github.com/nv-tlabs/GET3D
/? midjourney stable diffusion DALL-E "Imagen" comparison https://www.google.com/search?q=midjourney+stable+diffusion+...
"Google will label fake images created with its [Imagen] A.I" (2023) https://news.ycombinator.com/item?id=35896000
"Show HN: Polymath: Convert any music-library into a sample-library with ML" https://news.ycombinator.com/item?id=34782526 :
> Other cool #aiart things:
JavaScript state machines and statecharts
State machine: https://en.wikipedia.org/wiki/Finite-state_machine
https://docs.viewflow.io/bpmn/index.html :
> Unlike Finite state machine based workflow, BPMN allows parallel task execution, and suitable to model real person collaboration patterns.
BPMN: Business Process Model and Notation: https://en.wikipedia.org/wiki/Business_Process_Model_and_Not...
On the difference between the map and the territory.
> - Can TLA+ find side-channels (which bypass all software memory protection features other than encryption-in-RAM)?
From https://news.ycombinator.com/item?id=31617335 :
>> Can there still be side channel attacks in formally verified systems? Can e.g. TLA+ help with that at all?
From https://news.ycombinator.com/item?id=33563857 :
> - TLA+ Model checker https://en.wikipedia.org/wiki/TLA%2B#Model_checker :
>> The TLC model checker builds a finite state model of TLA+ specifications for checking invariance properties
This library implements Harel state charts, which are a superset of FSMs that allow for parallel and nested states.
Not sure what TLA+ has to do with either.
Rising number of lithium battery incidents on airplanes worry crew
"Researchers craft a fully edible battery" (2023) https://news.ycombinator.com/item?id=35885696
The Simplest Universal Turing Machine Is Proved
https://en.wikipedia.org/wiki/Church–Turing–Deutsch_principl... :
> In computer science and quantum physics, the Church–Turing–Deutsch principle (CTD principle) is a stronger, physical form of the Church–Turing thesis formulated by David Deutsch in 1985.[1] The principle states that a universal computing device can simulate every physical process.
Are qubits enough to simulate qudits and qutrits?
Run Llama 13B with a 6GB graphics card
From skimming, it looks like this approach requires CUDA and thus is Nvidia only.
Anyone have a recommended guide for AMD / Intel GPUs? I gather the 4 bit quantization is the special sauce for CUDA, but I’d guess there’d be something comparable for not-CUDA?
4-bit quantization is to reduce the amount of VRAM required to run the model. You can run it 100% on CPU if you don't have CUDA. I'm not aware of any AMD equivalent yet.
Looks like there are several projects that implement the CUDA interface for various other compute systems, e.g.:
https://github.com/ROCm-Developer-Tools/HIPIFY/blob/master/R...
https://github.com/hughperkins/coriander
I have zero experience with these, though.
"Democratizing AI with PyTorch Foundation and ROCm™ support for PyTorch" (2023) https://pytorch.org/blog/democratizing-ai-with-pytorch/ :
> AMD, along with key PyTorch codebase developers (including those at Meta AI), delivered a set of updates to the ROCm™ open software ecosystem that brings stable support for AMD Instinct™ accelerators as well as many Radeon™ GPUs. This now gives PyTorch developers the ability to build their next great AI solutions leveraging AMD GPU accelerators & ROCm. The support from PyTorch community in identifying gaps, prioritizing key updates, providing feedback for performance optimizing and supporting our journey from “Beta” to “Stable” was immensely helpful and we deeply appreciate the strong collaboration between the two teams at AMD and PyTorch. The move for ROCm support from “Beta” to “Stable” came in the PyTorch 1.12 release (June 2022)
> [...] PyTorch ecosystem libraries like TorchText (Text classification), TorchRec (libraries for recommender systems - RecSys), TorchVision (Computer Vision), TorchAudio (audio and signal processing) are fully supported since ROCm 5.1 and upstreamed with PyTorch 1.12.
> Key libraries provided with the ROCm software stack including MIOpen (Convolution models), RCCL (ROCm Collective Communications) and rocBLAS (BLAS for transformers) were further optimized to offer new potential efficiencies and higher performance.
https://news.ycombinator.com/item?id=34399633 :
>> AMD ROcm supports Pytorch, TensorFlow, MlOpen, rocBLAS on NVIDIA and AMD GPUs: https://rocmdocs.amd.com/en/latest/Deep_learning/Deep-learni...
https://github.com/intel/intel-extension-for-pytorch :
> Intel® Extension for PyTorch extends PyTorch with up-to-date features optimizations for an extra performance boost on Intel hardware. Optimizations take advantage of AVX-512 Vector Neural Network Instructions (AVX512 VNNI) and Intel® Advanced Matrix Extensions (Intel® AMX) on Intel CPUs as well as Intel Xe Matrix Extensions (XMX) AI engines on Intel discrete GPUs. Moreover, through PyTorch xpu device, Intel® Extension for PyTorch provides easy GPU acceleration for Intel discrete GPUs with PyTorch
https://pytorch.org/blog/celebrate-pytorch-2.0/ (2023) :
> As part of the PyTorch 2.0 compilation stack, TorchInductor CPU backend optimization brings notable performance improvements via graph compilation over the PyTorch eager mode.
> The TorchInductor CPU backend is sped up by leveraging the technologies from the Intel® Extension for PyTorch for Conv/GEMM ops with post-op fusion and weight prepacking, and PyTorch ATen CPU kernels for memory-bound ops with explicit vectorization on top of OpenMP-based thread parallelization
DLRS Deep Learning Reference Stack: https://intel.github.io/stacks/dlrs/index.html
Matter Raspberry Pi GPIO Commander – Turn Your Pi into a Matter Lighting Device
(Context: Matter is an application layer protocol to control smart home devices, bring them into your Wifi or Thread network, and they ought to work with Homekit/Google Home/Alexa/Smartthings/Home Assistant etc. at the same time)
I'm currently working on a Matter device myself and the overengineering is unreal.
The idea is great (standardized messages over IP, no internet required, open source SDK on Github) but if they had just used MQTT with some standardized JSON messages and two or three fixed onboarding flows it would have solved the same problems with 1% of the effort.
But now the project is absolutely massive, almost 1k pages for the base spec, 1M+ lines of (mostly C++) code for the core SDK, no stable or even well defined API, almost no documentation. Progress is glacial, unsuprisingly. Bug fixes do not seem to be backported to older release branches but there is no source code compatibility either, so the security story will be interesting for sure...
Most discussions seem to happen on an internal Discord server and Wiki for paying members only, so the project is not that open either.
I really hope that the entire thing gets scrapped and a simpler approach will rise from the ashes...
> I really hope that the entire thing gets scrapped and a simpler approach will rise from the ashes...
Do we need a new thing? ZigBee and mqtt already exist. Maybe something a little nicer over IP networks?
"Securing name resolution in the IoT: DNS over CoAP" https://news.ycombinator.com/item?id=32186286
From https://github.com/project-chip/connectedhomeip#architecture... :
> Matter aims to build a universal IPv6-based communication protocol for smart home devices. The protocol defines the application layer that will be deployed on devices and the different link layers to help maintain interoperability. The following diagram illustrates the normal operational mode of the stack:
> [...] It is built with market-proven technologies using Internet Protocol (IP) and is compatible with Thread and Wi-Fi network transports.
> Matter was developed by a Working Group within the Connectivity Standards Alliance (Alliance). This Working Group develops and promotes the adoption of the Matter standard, a royalty-free connectivity standard to increase compatibility among smart home products, with security as a fundamental design tenet. The vision that led major industry players to come together to build Matter is that smart connectivity should be simple, reliable, and interoperable.
> [...] The code examples show simple interactions, and are supported on multiple transports -- Wi-Fi and Thread -- starting with resource-constrained (i.e., memory, processing) silicon platforms to help ensure Matter’s scalability.
Would it make sense to have an mqtt transport for matter? Or a bridge; like homeassistant or similar?
https://github.com/home-assistant/core#featured-integrations lists mqtt
Is there already a good (security) comparison of e.g. http basic auth, x10, ZigBee, mqtt, matter?
Exploring the native use of 64-bit posit arithmetic in scientific computing
Unum (number format) > Posit (Type III Unum) https://en.wikipedia.org/wiki/Unum_(number_format) :
> Posits have superior accuracy in the range near one, where most computations occur. This makes it very attractive to the current trend in deep learning to minimise the number of bits used. It potentially helps any application to accelerate by enabling the use of fewer bits (since it has more fraction bits for accuracy) reducing network and memory bandwidth and power requirements.
> [...] Note: 32-bit posit is expected to be sufficient to solve almost all classes of applications [citation needed]
Researchers craft a fully edible battery
"Researchers find major storage capacity in metal-free aqueous batteries" https://www.inceptivemind.com/researchers-finds-major-storag... :
"The role of the electrolyte in non-conjugated radical polymers for metal-free aqueous energy storage electrodes." (2023) DOI: 10.1038/s41563-023-01518-z
"A sustainable battery with a biodegradable electrolyte made from crab shells" (2022) https://www.sciencedaily.com/releases/2022/09/220901135827.h... :
"A sustainable chitosan-zinc electrolyte for high-rate zinc-metal batteries" (2022) http://dx.doi.org/10.1016/j.matt.2022.07.015
Chitosan https://en.wikipedia.org/wiki/Chitosan :
> Chitosan /ˈkaɪtəsæn/ is a linear polysaccharide composed of randomly distributed β-(1→4)-linked D-glucosamine (deacetylated unit) and N-acetyl-D-glucosamine (acetylated unit). It is made by treating the chitin shells of shrimp and other crustaceans with an alkaline substance, such as sodium hydroxide.
Is there a sustainable way to produce Chitosan, which is an electrolyte?
https://westurner.github.io/hnlog/ Ctrl-F "anode" :
> Here's a discussion about the lower costs of hemp supercapacitors as compared with graphene super capacitors: https://news.ycombinator.com/item?id=16814022
"Why Salt Water may be the Future of Batteries" https://youtu.be/vm2hNNA4lvM ($5/kwh, energy density, desalinization too, graphene production too)
Flow battery > Other types > Membraneless https://en.wikipedia.org/wiki/Flow_battery#Other_types
MSG is the most misunderstood ingredient of the century. That’s finally changing
MSG: Monosodium Glutamate: https://en.wikipedia.org/wiki/Monosodium_glutamate
Glutamate–glutamine cycle: https://en.wikipedia.org/wiki/Glutamate%E2%80%93glutamine_cy...
"Multiple Mechanistically Distinct Modes of Endoc annabinoid Mobilization at Central Amygdala Glutamatergic Synapses" (2014) https://www.cell.com/neuron/fulltext/S0896-6273(14)00017-8
Glutamine https://en.wikipedia.org/wiki/Glutamine :
> Glutamine is synthesized by the enzyme glutamine synthetase from glutamate and ammonia. The most relevant glutamine-producing tissue is the muscle mass, accounting for about 90% of all glutamine synthesized. Glutamine is also released, in small amounts, by the lungs and brain. [20] Although the liver is capable of relevant glutamine synthesis, its role in glutamine metabolism is more regulatory than producing, since the liver takes up large amounts of glutamine derived from the gut. [7]
"Glutamine-to-glutamate ratio in the nucleus accumbens predicts effort-based motivated performance in humans" (2020) https://www.nature.com/articles/s41386-020-0760-6
> High glutamine-to-glutamate ratio predicts the ability to sustain motivation: The researchers found that individuals with a higher glutamine-to-glutamate ratio had a higher success rate and a lower perception of effort. https://neuroscienceschool.com/2020/10/11/how-to-sustain-mot...
ICYMI: You may appreciate this video...
Is cancer all about cellular energy from oxygen?
/? lactic acid atp: https://www.google.com/search?q=lactic+acid+atp
- "What is lactic acid?" https://www.livescience.com/what-is-lactic-acid
> Although blood lactate concentration does increase during intense exercise, the lactic acid molecule itself dissociates and the lactate is recycled and used to create more ATP.
> "Your body naturally metabolizes the lactic acid, clearing it out. The liver can take up some of the lactic acid molecules and convert them back to glucose for fuel," says Grover. "This conversion also reduces the acidity in the blood, thus removing some of the burning sensation. This is a natural process that occurs in the body. Things such as stretching, rolling, or walking will have little to no impact."
> The burning sensation you feel in your legs during a heavy workout probably isn't caused by lactic acid, but instead by tissue damage and inflammation.
> It’s also important to remember that lactate itself isn’t 'bad'. In fact, research in [Bioscience Horizons] suggests that lactate is beneficial to the body during and after exercise in numerous ways. For example, lactate can be used directly by the brain and heart for energy or converted into glucose in the liver or kidneys, which can then be used by nearly any cell in the body for energy.
- https://chem.libretexts.org/Courses/University_of_Kentucky/U... :
> Lactic Acid Fermentation: You may have not been aware that your muscle cells can ferment. Fermentation is the process of producing ATP in the absence of oxygen, through glycolysis alone. Recall that glycolysis breaks a glucose molecule into two pyruvate molecules, producing a net gain of two ATP and two NADH molecules. Lactic acid fermentation is the type of anaerobic respiration carried out by yogurt bacteria (Lactobacillus and others) and by your own muscle cells when you work them hard and fast.
Is that relevant to cancer and oxygen, IDK.
Hopefully NIRS + ultrasound ("HIFU" & modern waveguides) can inexpensively treat many more forms of cancer.
From https://news.ycombinator.com/item?id=35812168 :
> https://news.ycombinator.com/item?id=35617859 ... "A simple technique to overcome self-focusing, filamentation, supercontinuum generation, aberrations, depth dependence and waveguide interface roughness using fs laser processing" [w/ Dual Beams]
See this page fetch itself, byte by byte, over TLS
https://github.com/jawj/subtls (A proof-of-concept TypeScript TLS 1.3 client) is implemented with the SubtleCrypto API.
TIL about SubtleCrypto https://developer.mozilla.org/en-US/docs/Web/API/SubtleCrypt... :
> The SubtleCrypto interface of the Web Crypto API provides a number of low-level cryptographic functions. Access to the features of SubtleCrypto is obtained through the subtle property of the Crypto object you get from the crypto property.
decrypt()
deriveBits()
deriveKey()
digest()
encrypt()
exportKey()
generateKey()
importKey()
sign()
unwrapKey()
verify()
wrapkey()
Can SubtleCrypto accelerate any of the W3C Verifiable Credential Data Integrity 1.0 APIs? vc-data-integrity: https://w3c.github.io/vc-data-integrity/ ctrl-f "signature suite"> ISSUE: Avoid signature format proliferation by using text-based suite value The pattern that Data Integrity Signatures use presently leads to a proliferation in signature types and JSON-LD Contexts. This proliferation can be avoided without any loss of the security characteristics of tightly binding a cryptography suite version to one or more acceptable public keys. The following signature suites are currently being contemplated: eddsa-2022, nist-ecdsa-2022, koblitz-ecdsa-2022, rsa-2022, pgp-2022, bbs-2022, eascdsa-2022, ibsa-2022, and jws-2022.
But what about "Kyber, NTRU, {FIPS-140-3}? [TLS1.4/2.0?]" i.e. PQ Post-Quantum signature suites? Why don't those need to be URIs, too?
Loophole-free Bell inequality violation with superconducting circuits
"Loophole-free Bell test ‘Spooky action at a distance’, no cheating (tudelft.nl)" (2015) https://news.ycombinator.com/item?id=10430675
/? ./hnlog: "nonlocal", "unitarity":
From "The fundamental thermodynamic costs of communication" (2023) https://westurner.github.io/hnlog/#story-34765393 :
> Isn't there more entropy if we consider all possible nonlocal relations between bits; or, is which entropy metric independent of redundant coding schemes between points in spacetime?
From "New neural network architecture inspired by neural system of a worm" (2023) https://news.ycombinator.com/item?id=34809725 :
> Is there an information metric which expresses maximal nonlocal connectivity between bits in a bitstring; that takes all possible (nonlocal, discontiguous) paths into account?
> `n_nodes2` only describes all of the binary, pairwise possible relations between the bits or qubits in a bitstring?*
> "But what is a convolution" https://www.3blue1brown.com/lessons/convolutions
> Quantum discord: https://en.wikipedia.org/wiki/Quantum_discord
From "Formal methods only solve half my problems" (2022) https://westurner.github.io/hnlog/#comment-31625499 :
> It might be argued that side channels exist whenever there is a shared channel; which is always, because plenoptic functions, wave function, air gap, ram bus mhz, nonlocal entanglement
> Category:Side-channel attacks https://en.wikipedia.org/wiki/Category:Side-channel_attacks
/? ./hnlog: "entanglement" ... Quantum FlyTrap
Show HN: Mineo.app – Better Python Notebooks
Hello everyone,
I would like to introduce our startup to HN: Mineo.app. Mineo.app is a production-ready SaaS Python notebook that provides a complete environment for building your data applications: Dashboards, Reports, and Data Pipelines based on Python notebooks.
Key features:
* Superpowered jupyter-compatible Python notebooks with extra goodies like: version control, commenting support, custom docker images, etc... enhanced with no code components that allow to create beautiful dashboards and reports.
* Data Pipelines: Ability to schedule and run one or more notebooks.
* Integrated file system to manage your files and projects with detailed permissions and groups.
We have a freemium licensing model, so you can start using Mineo just by registering with your Github/Google/Microsoft account for free without a credit card. And it's free for educational purposes ;-)
Diego.
I think it was a little too early share but since you asked...
Crafting workflows out of notebooks is a really bad idea; an anti-pattern. If you want to go down the road of "workflows for data scientists" look at https://featurebyte.com/
Your messaging indicates you also want to appeal to business intelligence users, with their data exploration needs. In this space I give the crown to https://hex.tech/ How are you going to be different or better?
Also, you should fix your showcase so prospective users don't have to log in. I'll just nope out of your site.
Have you considered going open core? That would make me look more favorably at your product given your lack of indisputable product- and technical superiority.
And last but not least, I wish you success!
> Crafting workflows out of notebooks is a really bad idea; an anti-pattern. If you want to go down the road of "workflows for data scientists"
https://westurner.github.io/hnlog/ Ctrl-F "DVC" ( https://dvc.org/ ) , https://westurner.github.io/hnlog/#comment-24261118 "Ten Simple Rules for Reproducible Computational Research", “Ten Simple Rules for Creating a Good Data Management Plan”, PROV
pygwalker https://github.com/Kanaries/pygwalker :
> PyGWalker: Turn your pandas dataframe into a Tableau-style User Interface for visual analysis
"Generate code from GUI interactions; State restoration & Undo" https://github.com/Kanaries/pygwalker/issues/90
The Scientific Method is testing, so testing (tests, assertions, fixtures) should be core to any scientific workflow system.
- [ ] (It's not possible to run `!pytest` in a Jupyter notebook without installing an extension with JupyterLite in WASM onnly where there's not yet a terminal or even yet a slow-but-usable [cheerpx] webvm bridged to jupyter kernel WASM ~process-space.)
awesome-jupyter#testing: https://github.com/markusschanta/awesome-jupyter#testing
ml-tooling/best-of-jupyter lists papermill/papermill under "Interactive Widgets/Visualization" https://github.com/ml-tooling/best-of-jupyter#interactive-wi...
"Markdown based notebooks" would store files next to the .ipynb.md, which implies a need for an MHTML/ZIP-like archive (for report notebook artifacts produced by scientific workflow systems with provenance metadata); but W3C Web Bundles avoid modifying linked resources with new specs: https://github.com/jupyter/enhancement-proposals/pull/103#is...
Google will label fake images created with its A.I
> Google’s approach is to label the images when they come out of the AI system, instead of trying to determine whether they’re real later on. Google said Shutterstock and Midjourney would support this new markup approach. Google developer documentation says the markup will be able to categorize images as trained algorithmic media, which was made by an AI model; a composite image that was partially made with an AI model; or algorithmic media, which was created by a computer but isn’t based on training data.
Can it store at least: (1) the prompt; and (2) the model which purportedly were generated by a Turing robot with said markup specification? Is it schema.org JSON-LD?
It's IPTC: https://developers.google.com/search/docs/appearance/structu...
If IPTC-to-RDF i.e./e.g. schema:ImageObject (schema:CreativeWork > https://schema.org/ImageObject) mappings are complete, it would be possible to sign IPTC metadata with W3C Verifiable Credentials (and e.g. W3C DIDs) just like any other [JSON-LD,] RDF; but is there an IPTC schema extension for appending signatures, and/or is there an IPTC graph normalization step that generates equivalent output to a (web-standardized) JSON-LD normalization function?
/? IPTC jsonschema: https://github.com/ihsn/nada/blob/master/api-documentation/s...
/? IPTC schema.org RDFS
IPTC extension schema: https://exiv2.org/tags-xmp-iptcExt.html
[ Examples of input parameters & hyperparameters: from e.g. the screenshot in the README.md of stablediffusion-webui or text-generation-webui: https://github.com/AUTOMATIC1111/stable-diffusion-webui ]
How should input parameters and e.g. LLM model version & signed checksum and model hyperparameters be stored next to a generated CreativeWork? filename.png.meta.jsonld.json or similar?
If an LLM passes the Turing test ("The Imitation Game") - i.e. has output indistinguishable from a human's output - does that imply that it is not possible to stylometrically fingerprint its outputs without intentional watermarking?
https://en.wikipedia.org/wiki/Turing_test
Implicit in the Turing test is the entity doing the evaluation. It's quite possible that a human evaluator could be tricked, but a tool-assisted human, or an AI itself could not be. Or even just some humans could be better at not being tricked than others.
Tell HN: We should start to add “ai.txt” as we do for “robots.txt”
I started to add an ai.txt to my projects. The file is just a basic text file with some useful info about the website like what it is about, when was it published, the author, etc etc.
It can be great if the website somehow ends up in a training dataset (who knows), and it can be super helpful for AI website crawlers, instead of using thousands of tokens to know what your website is about, they can do it with just a few hundred.
Your HTML already has semantic meta elements like author and description you should be populating with info like that: https://developer.mozilla.org/en-US/docs/Learn/HTML/Introduc...
and also opengraph meta tags https://ogp.me/
And also schema.org: https://schema.org/
Thing > CreativeWork > WebSite https://schema.org/WebSite ... scroll down to "Examples" and click the "JSON-LD" and/or "RDFa" tabs. (And if there isn't an example then go to the schema.org/ URL of a superClassOf (rdfs:subClassOf) of the rdfs:Class or rdfs:Property; there are many markup examples for CreativeWork and subtypes).
Also: https://news.ycombinator.com/item?id=35891631
extruct is one way to parse linked data from HTML pages: https://github.com/scrapinghub/extruct
security.txt https://github.com/securitytxt/security-txt :
> security.txt provides a way for websites to define security policies. The security.txt file sets clear guidelines for security researchers on how to report security issues. security.txt is the equivalent of robots.txt, but for security issues.
Carbon.txt: https://github.com/thegreenwebfoundation/carbon.txt :
> A proposed convention for website owners and digital service providers to demonstrate that their digital infrastructure runs on green electricity.
"Work out how to make it discoverable - well-known, TXT records or root domains" https://github.com/thegreenwebfoundation/carbon.txt/issues/3... re: JSON-LD instead of txt, signed records with W3C Verifiable Credentials (and blockcerts/cert-verifier-js)
SPDX is a standard for specifying software licenses (and now SBOMs Software Bill of Materials, too) https://en.wikipedia.org/wiki/Software_Package_Data_Exchange
It would be transparent to disclose the SBOM in AI.txt or elsewhere.
How many parsers should be necessary for https://schema.org/CreativeWork https://schema.org/license metadata for resources with (Linked Data) URIs?
Having a security.txt doesn't stop security researchers asking "Do you have a bounty program?". We replied dozens already that such a file exist, it's not well enough known yet. On the other hand there are search engines crawling those and creating reports, which is nice.
JSON-LD or RDFa (RDF in HTML attributes) in at least the /index.html the HTML footer should be sufficient to indicate that there is structured linked data metadata for crawlers that then don't need an HTTP request to a .well-known URL /.well-known/ai_security_reproducibility_carbon.txt.jsonld.json
OSV is a new format for reporting security vulnerabilities like CVEs and an HTTP API for looking up CVEs from software component name and version. https://github.com/ossf/osv-schema
A number of tools integrate with OSV-schema data hosted by osv.dev: https://github.com/google/osv.dev#third-party-tools-and-inte... :
> We provide a Go based tool that will scan your dependencies, and check them against the OSV database for known vulnerabilities via the OSV API.
> Currently it is able to scan various lockfiles [ repo2docker REES config files like and requirements.txt, Pipfile lock, environment.yml, or a custom Dockerfile, ], debian docker containers, SPDX and CycloneDB SBOMs, and git repositories.
Language models can explain neurons in language models
LLMs are quickly going to be able to start explaining their own thought processes better than any human can explain their own. I wonder how many new words we will come up with to describe concepts (or "node-activating clusters of meaning") that the AI finds salient that we don't yet have a singular word for. Or, for that matter, how many of those concepts we will find meaningful at all. What will this teach us about ourselves?
First of all, our own explanations about ourselves and our behaviour are mostly lies, fabrications, hallucinations, faulty re-memorization, post hoc reasoning:
"In one well-known experiment, a split-brain patient’s left hemisphere was shown a picture of a chicken claw and his right hemisphere was shown a picture of a snow scene. The patient was asked to point to a card that was associated with the picture he just saw. With his left hand (controlled by his right hemisphere) he selected a shovel, which matched the snow scene. With his right hand (controlled by his left hemisphere) he selected a chicken, which matched the chicken claw. Next, the experimenter asked the patient why he selected each item. One would expect the speaking left hemisphere to explain why it chose the chicken but not why it chose the shovel, since the left hemisphere did not have access to information about the snow scene. Instead, the patient’s speaking left hemisphere replied, “Oh, that’s simple. The chicken claw goes with the chicken and you need a shovel to clean out the chicken shed”" [1]. Also [2] has an interesting hypothesis on split-brains: not two agents, but two streams of perception.
[1] 2014, "Divergent hemispheric reasoning strategies: reducing uncertainty versus resolving inconsistency", https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4204522
[2] 2017, "The Split-Brain phenomenon revisited: A single conscious agent with split perception", https://pure.uva.nl/ws/files/25987577/Split_Brain.pdf
Not supported by neuroimaging. Promoted without evidence or sufficient causal inference.
https://www.health.harvard.edu/blog/right-brainleft-brain-ri... :
> But, the evidence discounting the left/right brain concept is accumulating. According to a 2013 study from the University of Utah, brain scans demonstrate that activity is similar on both sides of the brain regardless of one's personality.
> They looked at the brain scans of more than 1,000 young people between the ages of 7 and 29 and divided different areas of the brain into 7,000 regions to determine whether one side of the brain was more active or connected than the other side. No evidence of "sidedness" was found. The authors concluded that the notion of some people being more left-brained or right-brained is more a figure of speech than an anatomically accurate description.
Here's wikipedia on the topic: "Lateralization of brain function" https://en.wikipedia.org/wiki/Lateralization_of_brain_functi...
Furthermore, "Neuropsychoanalysis" https://en.wikipedia.org/wiki/Neuropsychoanalysis
Neuropsychology: https://en.wikipedia.org/wiki/Neuropsychology
Personality psychology > ~Biophysiological: https://en.wikipedia.org/wiki/Personality_psychology
MBTI > Criticism: https://en.wikipedia.org/wiki/Myers%E2%80%93Briggs_Type_Indi...
Connectome: https://en.wikipedia.org/wiki/Connectome
This is not relevant to GP's comment. It has nothing to do with "are there fixed 'themes' that are operated in each hemisphere." It has to do with more generally, does the brain know what the brain is doing. The answer so far does not seem to be "yes."
Says who? There is actual evidence to support that our brain doesn't "know" what it is doing on a subconscious level? As far as I'm aware it's more that conscious humans don't understand how our brain works.
I think the correct statement is "so far the answer is we don't know"
The split brain experiments very very clearly indicate that different parts of the brain can independently conduct behavior and gain knowledge independently of other parts.
How or if this generalizes to healthy brains is not super clear, but it does actually provide a good explanatory model for all sorts of self-contradictory behavior (like addiction): the brain has many semi-independent “interests” that are jockeying for overall control of the organism’s behavior. These interests can be fully contradictory to each other.
Correct, ultimately we do not know. But it’s actually a different question than your rephrasing.
Given that functional localization varies widely from subject to subject per modern neuroimaging, how are split brain experiments more than crude attempts to confirm functional specialization (which is already confirmed without traumatically severing a corpus callosum) "hemispheric" or "lateral"?
Neuroimaging indicates high levels of redundancy and variance in spatiotemporal activation.
Studies of cortices and other tissues have already shown that much of the neural tissue of the brain is general purpose.
Why is executive functioning significantly but not exclusively in the tissue of the forebrain, the frontal lobes?
Because there’s a version of specialization that is, “different regions are specialized but they all seem to build consensus” and there’s a version that is “different regions are specialized and consensus does not seem to be necessary or potentially even usual or possible.”
These offer very different interpretations of cognition and behavior, and the split brain experiments point toward the latter.
Functional specialization > Major theories of the brain> Modularity or/and Distributive processing: https://en.wikipedia.org/wiki/Functional_specialization_(bra... :
> Modularity: [...] The difficulty with this theory is that in typical non-lesioned subjects, locations within the brain anatomy are similar but not completely identical. There is a strong defense for this inherent deficit in our ability to generalize when using functional localizing techniques (fMRI, PET etc.). To account for this problem, the coordinate-based Talairach and Tournoux stereotaxic system is widely used to compare subjects' results to a standard brain using an algorithm. Another solution using coordinates involves comparing brains using sulcal reference points. A slightly newer technique is to use functional landmarks, which combines sulcal and gyral landmarks (the groves and folds of the cortex) and then finding an area well known for its modularity such as the fusiform face area. This landmark area then serves to orient the researcher to the neighboring cortex. [7]
Is there a way to address the brain with space-filling curves around ~loci/landmarks? For brain2brain etc
FWIU, Markham's lab found that the brain is at max 11D in some places; But an electron wave model (in the time domain) may or must be sufficient according to psychoenergetics (Bearden)
> Distributive processing: [...] McIntosh's research suggests that human cognition involves interactions between the brain regions responsible for processes sensory information, such as vision, audition, and other mediating areas like the prefrontal cortex. McIntosh explains that modularity is mainly observed in sensory and motor systems, however, beyond these very receptors, modularity becomes "fuzzier" and you see the cross connections between systems increase.[33] He also illustrates that there is an overlapping of functional characteristics between the sensory and motor systems, where these regions are close to one another. These different neural interactions influence each other, where activity changes in one area influence other connected areas. With this, McIntosh suggest that if you only focus on activity in one area, you may miss the changes in other integrative areas.[33] Neural interactions can be measured using analysis of covariance in neuroimaging [...]
FWIU electrons are most appropriately modeled with Minkowski 4-space in the time-domain; (L^3)t
Neuroplasticity: https://en.wikipedia.org/wiki/Neuroplasticity :
> The adult brain is not entirely "hard-wired" with fixed neuronal circuits. There are many instances of cortical and subcortical rewiring of neuronal circuits in response to training as well as in response to injury.
> There is ample evidence [53] for the active, experience-dependent re-organization of the synaptic networks of the brain involving multiple inter-related structures including the cerebral cortex.[54] The specific details of how this process occurs at the molecular and ultrastructural levels are topics of active neuroscience research. The way experience can influence the synaptic organization of the brain is also the basis for a number of theories of brain function
"Representational drift: Emerging theories for continual learning and experimental future directions" (2022) https://www.sciencedirect.com/science/article/pii/S095943882... :
> Recent work has revealed that the neural activity patterns correlated with sensation, cognition, and action often are not stable and instead undergo large scale changes over days and weeks—a phenomenon called representational drift. Here, we highlight recent observations of drift, how drift is unlikely to be explained by experimental confounds, and how the brain can likely compensate for drift to allow stable computation. We propose that drift might have important roles in neural computation to allow continual learning, both for separating and relating memories that occur at distinct times. Finally, we present an outlook on future experimental directions that are needed to further characterize drift and to test emerging theories for drift's role in computation.
So, to run the same [fMRI, NIRS,] stimulus response activation observation/burn-in again weeks or months later with the same subjects is likely necessary given Representational drift.
IPyflow: Reactive Python Notebooks in Jupyter(Lab)
That’s a pretty nice idea. The problem of knowing what state has been invalidated often drives me away from using a notebook. So it is nice to see this solved.
I often though we would benefit from having some kind of shell, only a mix between ipython qtconsole et jupyter.
Not an editor like jupyter, rather a shell with a REPL flow. But each prompt is like a jupyter cell, and the whole history is saved in a file.
But if you don't create a file, it should work as well. One of the annoying things about jupyter is that you can use it without file on disk unlike ipython shell.
jupyter_console is the IPython REPL for non-ipykernel jupyter kernels.
This magic command logs IPython REPL input and output to a file:
%logstart -o example.log.py
Machine Learning Containers Are Bloated and Vulnerable
https://repo2docker.readthedocs.io/en/latest/ :
> jupyter-repo2docker is a tool to build, run, and push Docker images from source code repositories.
> repo2docker fetches a repository (from GitHub, GitLab, Zenodo, Figshare, Dataverse installations, a Git repository or a local directory) and builds a container image in which the code can be executed. The image build process is based on the configuration files found in the repository.
> repo2docker can be used to explore a repository locally by building and executing the constructed image of the repository, or as a means of building images that are pushed to a Docker registry.
> repo2docker is the tool used by BinderHub to build images on demand
There are maintenance advantages and longer time to patch with kitchen-sink ML containers like kaggle/docker-python because it takes work to entirely bump all of the versions in the requirements specification files (and run integration tests to make sure code still runs after upgrading everything or one thing at a time).
What's best practice for including a sizeable dataset in a container (that's been recently re-) built with repo2docker?
Health advisory on social media use in adolescence
This recent article suggests other factors might be more important to adolescent mental health:
Time spent on social media among the least influential factors in adolescent mental health: preliminary results from a panel network analysis https://www.nature.com/articles/s44220-023-00063-7
> There is growing concern about the role of social media use in the documented increase of adolescent mental health difficulties. However, the current evidence remains complex and inconclusive. While increasing research on this area of work has allowed for notable progress, the impact of social media use within the complex systems of adolescent mental health and development is yet to be examined. The current study addresses this conceptual and methodological oversight by applying a panel network analysis to explore the role of social media on key interacting systems of mental health, wellbeing and social life of 12,041 UK adolescents. Here we find that, across time, estimated time spent interacting with social media predicts concentration problems in female participants. However, of the factors included in the current network, social media use is one of the least influential factors of adolescent mental health, with others (for example, bullying, lack of family support and school work dissatisfaction) exhibiting stronger associations. Our findings provide an important exploratory first step in mapping out complex relationships between social media use and key developmental systems and highlight the need for social policy initiatives that focus on the home and school environment to foster resilience.
> Time spent on social media among the least influential factors in adolescent mental health
But it's the one that's easy to assign blame to: it's the evil (sometimes foreign) big tech companies fault.
Increasingly poor economics prospects, environmental crisis and over-competitive society are much tougher issues to crack, and perhaps, in the case of housing for instance, certain demographics would prefer to blame the "evil screens" rather than their own generation's behavior over the years...
although I believe the big tech companies are doing their very best to capture attention at all costs, it is a good question to ask why people are in a position to spend so much time on social media and are doing so in place of other activities. For example, the lack of public spaces where one can spend time without spending money, relatively recent parenting practices and legal obstacles to having unsupervised time outdoors, etc.
I can see it in myself and I suspect others feel similarly. And I suspect kids feel it even more strongly but that's just an assumption of mine. The way "big tech companies are doing their very best to capture attention at all costs" is by designing their apps such that it's physically addicting. Once you're addicted it's hard to do the other things you mentioned because they don't release the same chemicals in your brain.
Would it be abusive or advisable to adapt edtech offerings in light of social media and slot machines' UX user experience findings?
While they should never appease students, can't infotech and edtech learn how to keep their attention, too?
Perhaps prompt engineering can help to create engaging educational content with substantive progress metrics?
"Build a game in JS (like game category XYZ) to teach quantum entropy to beginners"
And then what prompt additions could help to social media-ify the game?
How should social media reinforce human communication behaviors with or without the stated age of the user? Should there be a "D- because that's harassment" panda video to reinforce? Which presidential role models' communication styles should AI emulate?
I find it sad to consider that the most impactful thing to do to improve children's lives would be to ban them from social media due to their age; though, for the record, e.g. Facebook did originally require a .edu email address at an approving institution.
Hopefully, Khanmigo and similar AI edtech offerings will be more engaging than preferentially reviewing unacceptable abuse online; but kids and people still need to learn to interact respectfully online in order to succeed.
Reactivating Dormant Cells in the Retina Brings New Hope for Vision Regeneration
"Direct neuronal reprogramming by temporal identity factors" (2023) https://www.pnas.org/doi/10.1073/pnas.2122168120#abstract :
> Abstract: Temporal identity factors are sufficient to reprogram developmental competence of neural progenitors and shift cell fate output, but whether they can also reprogram the identity of terminally differentiated cells is unknown. To address this question, we designed a conditional gene expression system that allows rapid screening of potential reprogramming factors in mouse retinal glial cells combined with genetic lineage tracing. Using this assay, we found that coexpression of the early temporal identity transcription factors Ikzf1 and Ikzf4 is sufficient to directly convert Müller glial (MG) cells into cells that translocate to the outer nuclear layer (ONL), where photoreceptor cells normally reside. We name these “induced ONL (iONL)” cells. Using genetic lineage tracing, histological, immunohistochemical, and single-cell transcriptome and multiome analyses, we show that expression of Ikzf1/4 in MG in vivo, without retinal injury, mostly generates iONL cells that share molecular characteristics with bipolar cells, although a fraction of them stain for Rxrg, a cone photoreceptor marker. Furthermore, we show that coexpression of Ikzf1 and Ikzf4 can reprogram mouse embryonic fibroblasts to induced neurons in culture by rapidly remodeling chromatin and activating a neuronal gene expression program. This work uncovers general neuronal reprogramming properties for temporal identity factors in terminally differentiated cells.
Is it possible to produce or convert Müller glial cells with Nanotransfection (stroma reprogramming), too?
From https://news.ycombinator.com/item?id=33127646 https://westurner.github.io/hnlog/#comment-33129531 re: "Retinoid restores eye-specific brain responses in mice with retinal degeneration" (2022) :
> Null hypothesis: A Nanotransfection (vasculogenic stromal reprogramming) intervention would not result in significant retinal or corneal regrowth
> ... With or without: a nerve growth factor, e.g. fluoxetine to induce plasticity in the adult visual cortex, combination therapy with cultured conjunctival IPS, laser mechanical scar tissue evisceration and removal, local anesthesia, robotic support, Retinoid
> Nanotransfection: https://en.wikipedia.org/wiki/Tissue_nanotransfection :
>> Most reprogramming methods have a heavy reliance on viral transfection. [22][23] TNT allows for implementation of a non-viral approach which is able to overcome issues of capsid size, increase safety, and increase deterministic reprogramming
Ask HN: Did anyone ever create GitLaw?
A thread from 11 years ago[0] proposed a "A GitHub for Laws and Legal Documents". Did anyone ever develop something similar to this? Even without the ability to propose changes, being able to view a git history of passed bills would be useful for anyone who would want to analyze the changes over time.
If a tool has been built, please share it. If not, what are the major blockers to at least building the history / diff tool?
[0]: https://news.ycombinator.com/item?id=3967921
Various localities upload their statutes to a git repository hosted as a GitHub project; but I'm not aware of any using Pull Request (PR) workflows to propose or debate legislative bills (or to enter or link to case law precedent for review).
The US Library of Congress operates the THOMAS system for legislative bills. How could THOMAS be improved; in order to support evidence-based policy; in order to improve democracy in democratic republics like the US? https://en.wikipedia.org/wiki/THOMAS
How do state systems for legislative bills document workflows and beyond compare to THOMAS, the US federal Congress system? Why do we need 50+1 independent [open source?] software applications; could states work together on such essential systems?
IIRC, LOC invested in improving THOMAS with: markup language to typographically-readable HTML support?
It may be helpful to send an email to your state with the regex regular expression pattern necessary to make online statute section and subsection references a href links instead of non-clickable string. TIL the section sign ("§") is older than the hyperlink; https://en.wikipedia.org/wiki/Section_sign
Parliamentary informatics: https://en.wikipedia.org/wiki/Parliamentary_informatics
FWIU, diffing documents is a solved problem.
How could pull request workflows be improved in order to support legislative and post-legislative workflows (like controlling for whether a funded plan actually has the proscribed impact)?
High School and College Policy debate (CX; Cross Examination debate) enjoy equal rights to 'fiat'. IRL, we must control for whether the plan and funding have the predicted impact.
E.g. GitHub and GitLab have various features that could support legislative workflows: code owners, multiple approvals required to merge, emoji reactions on Issues and Pull Requests and comments therein such that you don't have to say why you voted a particular way, GPG-signatures required for commit and merge, 2FA.
FWIU, the Aragorn / district0x projects have some of the first DLT-based (Blockchain) systems for democracy; all of the rest of us trust one party to share root access and multi-site backup responsibilities and have insufficient DDOS protections.
There was a discussion on HN where the real blocker is everyone uses Microsoft Word, and all the related tools on it, and no one was going to change their workflow.
Any such legislative solution may involve meeting the users where they are, first. Understanding the whole lifecycle, so to speak.
Show HN: ReRender AI - Realistic Architectural Renders for AutoCAD/Blender Users
Different but the same problem: "Generate a heat sink heat exchanger with maximum efficiency, shaped like a passive solar home"
TIL from off-gridders and homesteaders about passive solar design so that the air moves through the home without HVAC (compared with high rise buildings where it is necessary to pump water up like a gravitational potential water tower.)
Does ReRender AI have features for sustainable architecture?
Prompt: "Design a passive solar high-rise building with maximal energy storage and production"
Notes from "Zero energy ready homes are coming" (2023) https://news.ycombinator.com/item?id=35064493
ReRender AI focuses on rendering, not sustainable design features. Architects must incorporate sustainability themselves. However, the AI can visually represent a passive solar high-rise if provided with the necessary design elements.
Exciton Fission Breakthrough Could Revolutionize Photovoltaic Solar Cells
> Researchers have resolved the mechanism of exciton fission, which could increase solar-to-electricity efficiency by one-third, potentially revolutionizing photovoltaic technology.
“Orbital-resolved observation of singlet fission” by Alexander Neef, Samuel Beaulieu, Sebastian Hammer, Shuo Dong, Julian Maklar, Tommaso Pincelli, R. Patrick Xian, Martin Wolf, Laurenz Rettig, Jens Pflaum and Ralph Ernstorfer, 12 April 2023, Nature. DOI: 10.1038/s41586-023-05814-1 : https://www.nature.com/articles/s41586-023-05814-1 :
> Abstract: Singlet fission may boost photovoltaic efficiency by transforming a singlet exciton into two triplet excitons and thereby doubling the number of excited charge carriers. The primary step of singlet fission is the ultrafast creation of the correlated triplet pair. Whereas several mechanisms have been proposed to explain this step, none has emerged as a consensus. The challenge lies in tracking the transient excitonic states. Here we use time- and angle-resolved photoemission spectroscopy to observe the primary step of singlet fission in crystalline pentacene. Our results indicate a charge-transfer mediated mechanism with a hybridization of Frenkel and charge-transfer states in the lowest bright singlet exciton. We gained intimate knowledge about the localization and the orbital character of the exciton wave functions recorded in momentum maps. This allowed us to directly compare the localization of singlet and bitriplet excitons and decompose energetically overlapping states on the basis of their orbital character. Orbital- and localization-resolved many-body dynamics promise deep insights into the mechanics governing molecular systems and topological materials
Energy conversion efficiency: https://en.wikipedia.org/wiki/Energy_conversion_efficiency
Multiple exciton generation: https://en.wikipedia.org/wiki/Multiple_exciton_generation
Latex users are slower than Word users and make more errors (2014)
And neither typesetting activity results in reusable LinkedData; because it's still not possible to publish Linked Data (or indeed data and its schema) with LaTeX, or Word, or PDF.
ScholarlyArticles' most relevant purpose is to link to the schema:Dataset that the presented analysis is predicated upon.
ScholarlyArticle authors should consider the value of data reuse and (linked data) schema in choosing a typesetting and publishing format.
Is there a better way to publish Linked Data with existing tools like LaTeX, PDF, or Word? Which support CSVW? Which support RDF/RDFa/JSON-LD?
How could authors express experimental study controls with URIs; with qualified typed edges to the data (and a cryptographic signature from an IRB and the ScholarlyArticle's authors).
> Is there a better way to publish Linked Data with existing tools like LaTeX, PDF, or Word? Which support CSVW? Which support RDF/RDFa/JSON-LD?
- [ ] ~"How to publish Linked Data with Jupyter notebooks" #todo #LinkedReproducibility #LinkedResearch
- [ ] westurner/nbmeta, jupyter*/?: Linked Data result object for notebooks; with a _repr_mimebundle_() and application/ld+json
- [ ] python 3.x+ grammar: de-restrict use of the walrus assignment operator := so that users can assign to the LD dict result object and implicitly IPython.display.display() it because walrus assignment returns the object assigned (whereas normal = assignment does not return a value)
- [ ] westurner/nbmeta?, JLab, jupyter-book: JSON-LD Playground widget to display and reframe JSON-LD (& maybe YAML-LD, too)
- [ ] jupyterlab: schema.org notebook level bibliographic schema.org/CreativeWork metadata
- [ ] jupyterlab: JS/TypeScript: JSON-LD/YAML-LD editor widget also built on codemirror like JLab or a different existing RDFJS library or?
- [ ] rdflib, jupyter nb, sphinx, pygments: add YAML-LD syntax support (now that there's a W3C spec)
- [ ] MyST markdown, sphinx, docutils: YAML-LD in MyST [in: code-fence attr syntax,] in order to add Linked Data to MyST Markdown documents and thus Jupyter Notebooks and Jupyter Books
- [ ] hypothesis/h, sphinx-comments: Linked Data Annotations (within nested Markdown, like hypothesis' W3C Web Annotations support); cc re: 'MyST-LD' Markdown: quoting "@id", "@context" in YAML-LD
- [ ] atomspace-like TrustValues with RDFstar/SPARQLstar (in YAML-LD because that implies JSON-LD)
- [ ] dvc, GitHub Actions, GitLab CI, Gitea Actions,: how to add PROV RDF Linked Data metadata to workflows like DVC.org's & container+command-in-YAML approach
- [ ] Microsoft/excel, Microsoft/VSCode: How to CSVW and PROV with spreadsheets and or code ; see also Excel speed-running competitons
- [ ] REQ: jupyter/rtc+linkeddata/dokieli: howto integrate Jupyter notebooks with the list of specs in the dokieli README
- [ ] njupyter/nbformat#?: a post- .ipynb notebook + resources spec? W3C Web Bundles have advantages including: you don't have to rewrite URLs in content saved offline like MHTML (mv $1.mhtml $1.mhtml.zip)
- [ ] JupyterLab, VSCode: CoCalc - which supports Time Travel version control over notebooks, LaTeX docs, - added a TODO: ~not_editable code cell metadata item; which is more like lab notebooks in pen; but there's not GUI support in JLab or other Jupyter notebook nbformat implementations yet
- [ ] Jupyter/nbformat, ipywidgets, jupyterlab,: How to save widget state: https://github.com/jupyter-widgets/ipywidgets/issues/2465
> How could authors express experimental study controls with URIs; with qualified typed edges to the data (and a cryptographic signature from an IRB and the ScholarlyArticle's authors).
Verifiable Credentials Data Model v1.1: https://www.w3.org/TR/vc-data-model/
Verifiable Credential Data Integrity 1.0: Securing the Integrity of Verifiable Credential Data: https://www.w3.org/TR/vc-data-integrity/ :
> Abstract: This specification describes mechanisms for ensuring the authenticity and integrity of Verifiable Credentials and similar types of constrained digital documents using cryptography, especially through the use of digital signatures and related mathematical proofs.
Because data quality, data reuse, code reuse, web standard specifications, structured data outputs with provenance metadata for partially-automated metaanalyses,; https://5stardata.info/
The skills gap for Fortran looms large in HPC
> The good news for some HPC simulations and models, both inside of the National Nuclear Security Administration program at the DOE and in the HPC community at large, is that many large-scale physics codes have been rewritten or coded from scratch in C++
When C++ one day goes out of fashion, like Fortran already has, this will turn out to be bad news. Fortran is not a big language, and any programmer can learn Fortran in a matter of weeks. But if you'd need to teach people C++ just so that they'd be able to maintain a legacy code base, that will be considerably more difficult than with Fortran.
Probably in a thousand years nobody will be writing C++. In one year, there will still be a helluva lot of C++, including new projects.
Point being, the premise of your argument isn't granted. Is C++ going away soon? Or is it holding steady and even growing? How can someone objectively know either way?
> Is C++ going away soon? Or is it holding steady and even growing? How can someone objectively know either way?
No. It's not going away anytime soon, but you can look at high-performance languages that are gaining in popularity, and that may point to change in the future. TIOBE isn't perfect, but it's a good indicator of interest, and Rust and Go seem to continue to gain popularity.
Interestingly, and related to this article, Fortran has moved up from #31 to #20 on TIOBE's index, which really does speak to the importance of a math-optimized, high performance, compiled language. Another interesting change is MATLAB moving up from #20 to #14.
TIOBE index is rivaled by the annual Stack Overflow Developer Survey in some regards https://survey.stackoverflow.co/2022/#technology-most-popula... :
> Most popular technologies This year, we're comparing the popular technologies across three different groups: All respondents, Professional Developers, and those that are learning to code.
Most popular technologies > Programming, scripting, and markup languages
The Top 500 Green500 (and the TechEmpower Web Framework benchmarks) are also great resources for estimating what people did this past year; what are the "BigE" of our models in terms of water, kwh of [directly or PPA-offset] sourced clean energy.
The Linux Kernel Module Programming Guide
Ctrl-F "rust"
https://rust-for-linux.com/ links to LWN articles at https://lwn.net/Kernel/Index/#Development_tools-Rust that suggest that only basic modules are yet possible with the rust support in Linux kernels 6.2 and 6.3.
Rust-for-linux links to the Android binder module though: https://rust-for-linux.com/Android-Binder-Driver.html :
> Android Binder Driver: This project is an effort to rewrite Android's Binder kernel driver in Rust.
> Motivation: Binder is one of the most security and performance critical components of Android. Android isolates apps from each other and the system by assigning each app a unique user ID (UID). This is called "application sandboxing", and is a fundamental tenet of the Android Platform Security Model.
> The majority of inter-process communication (IPC) on Android goes through Binder. Thus, memory unsafety vulnerabilities are especially critical when they happen in the Binder driver
... "Rust in the Linux kernel" (2021) https://security.googleblog.com/2021/04/rust-in-linux-kernel... :
> [...] We also need designs that allow code in the two languages to interact with each other: we're particularly interested in safe, zero-cost abstractions that allow Rust code to use kernel functionality written in C, and how to implement functionality in idiomatic Rust that can be called seamlessly from the C portions of the kernel.
> Since Rust is a new language for the kernel, we also have the opportunity to enforce best practices in terms of documentation and uniformity. For example, we have specific machine-checked requirements around the usage of unsafe code: for every unsafe function, the developer must document the requirements that need to be satisfied by callers to ensure that its usage is safe; additionally, for every call to unsafe functions (or usage of unsafe constructs like dereferencing a raw pointer), the developer must document the justification for why it is safe to do so.
> We'll now show how such a driver would be implemented in Rust, contrasting it with a C implementation. [...]
Is this the source for the rust port of the Android binder kernel module?: https://android.googlesource.com/platform/frameworks/native/...
This guide with unsafe rust that calls into the C, and then with next gen much safer rust right next to it would be a helpful resource too.
What of the post-docker container support (with userspaces also written in go) should be cloned to rust first?
What are some good examples of non-trivial Linux kernel modules written in Rust?
On Android the Linux kernel is its own thing, and after Project Treble, it follows a microkernel like approach to drivers, where standard Linux drivers are considered "legacy" since Android 8.
https://source.android.com/docs/core/architecture/hal
Don't use how Linux kernel does things on Android as understanding from how upstream works.
In your opinion, do you think that the microkernel approach is more secure? (Should processes run as separate users with separate SELinux contexts like Android 4.4+)
Why do you think that the Android binder module rust implementation is listed as an example of a Rust for Linux kernel module on the site?
"Android AOSP Can Boot Off Mainline Linux 5.9 With Just One Patch" (2020) https://www.phoronix.com/news/Android-AOSP-Close-Linux-5.9 :
> The Android open-source project "AOSP" with its latest code is very close to being able to boot off the mainline Linux kernel when assuming the device drivers are all upstream.
Other distros support kmods and akmods: https://www.reddit.com/r/PINE64official/comments/ijfbgl/comm... :
> How kmod / akmod // DKMS work is something that the community is maybe not real familiar with.
Of course microkernel approach is more secure, if a driver gets p0wned, corrupts data structures, or plain keeps crashing, it doesn't take the whole kernel with it.
Naturally the issue might be as bad that the whole stack can't recover from, but still much better than corrupting the kernel.
One of the SecDevOps guidelines when hardening servers is that every process should have its own user, yes.
Kmods and akmods run in kernel memory space and aren't ABI stable.
Binder is not a HAL, since binder is how HALs communicate with each other. It's an actual proper Linux driver. The C version is in the upstream kernel as a module you can enable when building the kernel.
Waydroid (Android in containers) requires binder and optionally ashmem, though ashmem is not required anymore because memfd works with vanilla kernel: https://wiki.archlinux.org/title/Waydroid#Kernel_Modules
There is a Google Play Certification process for waydroid devices: https://docs.waydro.id/faq/google-play-certification
(That manual provisioning step is not necessary for e.g. optional widevine DRM on other OSes)
xPrize Wildfire – $11M Prize Competition
On one hand this effort sounds like a good way to generate some interesting concepts. On the other hand, as a long-time resident of a rural, wildfire-prone area, it's also important to educate the public and politicians that there is a large body of proven land management best practices that can already significantly improve prevention, control and mitigation. While we can always do better, we often aren't sufficiently utilizing the tools and techniques we already have.
The issues are more often political, economic and systemic than they are not knowing effective ways to further reduce risks and harm. Things like properly managed controlled burns are under-utilized because they are politically controversial. Enacting and enforcing codes mandating land management around private structures are politically unpopular. Out where I live, the local fire rangers just point-blank tell property owners "If you don't clear out all the brush and downed trees within a couple hundred feet of your structures, we're defending your house after the others that meet code" but no politician is willing to tell voters that. Worse, land owners continue to get permits issued to construct permanent dwellings in isolated locations which are extremely difficult to defend. If they want to build there, I say let them but also issue fair warning if they choose to proceed, they are on their own in the event of fire. Volunteer firefighters shouldn't need to risk their lives to defend houses which should never have been built in inaccessible, indefensible locations.
Wildfire suppression: https://en.wikipedia.org/wiki/Wildfire_suppression#Tactics
History of wildfire suppression in the United States: https://en.wikipedia.org/wiki/History_of_wildfire_suppressio...
Controlled burn: https://en.wikipedia.org/wiki/Controlled_burn
> A controlled or prescribed burn, also known as hazard reduction burning, [1] backfire, swailing, or a burn-off, [2] is a fire set intentionally for purposes of forest management, fire suppression, farming, prairie restoration or greenhouse gas abatement. A controlled burn may also refer to the intentional burning of slash and fuels through burn piles. [3] Fire is a natural part of both forest and grassland ecology and controlled fire can be a tool for foresters.
Hydrogel: https://en.wikipedia.org/wiki/Hydrogel
Aerogel: https://en.wikipedia.org/wiki/Aerogel
Water-based batteries have less fire risk FWIU
"1,000% Difference: Major Storage Capacity in Water-Based Batteries Found" (2023) https://scitechdaily.com/1000-difference-major-storage-capac...
> The metal-free water-based batteries are unique from those that utilize cobalt in their lithium-ion form. The research group’s focus on this type of battery stems from a desire for greater control over the domestic supply chain as cobalt and lithium are commonly sourced from outside the country. Additionally, the batteries’ safer chemistry could prevent fires.
That's a good list of useful links. From my experience as a home owner in a high-risk area, I'd add some of the things which can make a big difference are often surprisingly simple. In addition to managing your own land properly, make sure you have a large enough water tank to support firefighters in protecting your land. We got a well tank five times bigger than our typical usage needs. Another key thing is to put your tank next to where a fire truck can access it easily and get the proper adapter installed so the fire truck can direct connect to your tank (it's not included in most).
We also widened the access road on our property to support larger fire trucks and dozers. Obviously, fire-resistant construction materials and techniques are a huge help and not all require extensive retrofitting. There are spark resistant attic air access covers which are easy to install. We needed to re-roof and remodel anyway so we went with fire-resistant roofing and concrete siding. It's also important to have non-cellular local communications because the cell towers often go out first (our community uses 2-meter handsets for emergency comms). Doing those things along with the fact we've cleared all trees and brush out beyond 300 feet and only have metal out-buildings caused our local fire chief to comment that if everyone did what we've done, protecting the whole region would be three times easier. We also have enough generator power (on auto-failover) to pump our own well water to wet down our house and surroundings. With that our house has a really strong probability of surviving with no external help even if our property is directly hit by a major wildfire.
Edit to Add: I forgot to mention another key thing. If you live in a place like this, there's often only a single narrow public road serving your property and others nearby. Make sure you have transportation that'll let you get out cross-country if the road is blocked by either fire or heavy equipment. We have AWD ATVs fueled and always ready to go (plus they are handy around big properties like this). Between us and our neighbors we have a plan to get everyone on our road out without outside help - including a few elderly folk who will need assistance. Firefighters needing to rescue unprepared residents diverts vital resources from fighting the fire and clogs roads - which just makes everything else worse.
Great comments and advice. My neighborhood burned in the #GlassFire [0]. While our property was severely damaged, our home survived, due to many of the hardening techniques you mentioned and great work by CalFire. We had evacuated the night before and were staying in a local hotel watching the flames advance on our house via our security cameras. (Redundant net connections and a generator kept everything online). As our deck started to burn I called 911 to report it, figuring nothing would happen. About 20 minutes later a fireman walked into the frame of the camera and 2 minutes later the deck was out and the house was saved. I was cheering like I won the Super Bowl. So I really appreciate all the hardening I did, on-site water and CalFire.
When we rebuilt the exterior, I installed a set of sprinklers that ring the house and cover the roof/deck. They are not designed to fight the fire, instead I can activate them before a fire arrives to get things good and wet. This perimeter reduces the available fuels, reduces the heat load on the structure and reduces the risk of ember cast. It was a fun project with an ESP controller to sequence the valves and provide remote control.
Over the last few years, the Alert Wildfire Camera system, now over 700 cameras, has been a valuable resource. [1] Early detection of the fire, before it grows to extreme size has made a big difference. In the North Bay of California last year, this early detection and CalFire having nearby Air Attack resources on standby kept many small fires from becoming large ones in this area.
Another great resource is the Watch Duty app. [2] This is a non-profit, volunteer, but extremely professional service. They started in Northern California and are rapidly expanding. It is the go-to resource for Wildfire related information in the area.
Lastly, if anyone is looking at the X prize for satellite detection, I’d spend some time researching MODIS/VIIRS data products from NASA. [3]. A good starting point in some of the challenges of wildfire detection from space.
[0]. Fire tornado outside my window. https://youtu.be/1tZjWqh3-EU
[1]. https://www.alertwildfire.org/
TIL about building materials and fire hazard:
Fire resistance rating: https://en.wikipedia.org/wiki/Fire-resistance_rating
R-value: https://en.wikipedia.org/wiki/R-value_(insulation)
Here's a prompt to help with selecting sustainable building materials: """Generate a JSON-LD document of data with citations comparing building materials by: R-value, Fire Resistance Rating, typical structural integrity longevity in terms of years, VOCs at year 1 and year 5, and Sustainability factors like watt-hours and carbon and other pollutive outputs to produce, transport, distribute, install, and maintain . Include hempcrete (hemp and lime, which can be made from algae), rammed earth with and without shredded hemp hurds, wood, structural concrete, bamboo, and other emerging sustainable building materials.
Demonstrate loading the JSON data into a Pandas DataFrame and how to sort by multiple columns (with references to the docs for the tools used), and how to load the data into pygwalker."""
> a set of sprinklers that ring the house and cover the roof/deck. They are not designed to fight the fire, instead I can activate them before a fire arrives to get things good and wet. This perimeter reduces the available fuels, reduces the heat load on the structure and reduces the risk of ember cast. It was a fun project with an ESP controller to sequence the valves and provide remote control.
I searched a bit and couldn't find any smart home integrations that automatically hopefully turn the lawn and garden sprinklers on when a fire alarm goes off.
Are there any good reasons to not have that be a standard default feature if both smoke detectors and sprinklers are connected to a smart home system?
Coming from an operations perspective, you want to look at the downstream consequences of automation which depend on the details of the installed system. E.g. from this thread: automation might drain water from the 5x oversized well tank before firefighters arrive, seems to me that whether or not that is preferable depends on the details.
> Hydrogel
From https://twitter.com/westurner/status/1572664456210948104 :
>> What about #CO² -based #hydrogels for fire fighting?
>> /? Hydrogel firefighting (2022) https://www.google.com/search?q=hydrogel+firefighting […]
>> IDEA: How to make #hydrogels (and #aerogels) from mostly just {air*, CO², algae, shade, and sunshine,} [on earth, and eventually in space,]?
> Aerogel
From https://twitter.com/westurner/status/1600820322567041024 :
> Problem: #airogel made of CO² is an excellent insulator that's useful for many applications; but it needs structure, so foam+airogel but that requires smelly foam
> Possible solution: Cause structure to form in the airogel.
Backpack shoulder straps on e.g. Jansport backpacks have a geometric rubber mesh that's visible through a plastic window.
## Possible methods of causing structure to form in aerogel
- EM Hz: literally play EM waves (and possibly deliberately inverse convolutions) at the {#airogel,} production process
- Titratation
- Centrifugation
- "Volt grid": apply volts/amps/Hz patterns with a probe array
- Thermal/photonic bath 3d printing
- Pour a lattice-like lens as large as the aerogel sections and allow solar to cause it to slowly congeal to a more structural form in advantageous shapes
- Die-casting, pressure-injection molding
> Water-based batteries have less fire risk FWIU
CAES Compressed Air Energy Storage and Gravitational potential energy storage also have less fire risk then Lithium Ion batteries.
FWIU we already have enough abandoned mines in the world to do all of our energy storage needs?
Could CAES tanks filled with air+CO² help suppress wildfire?
> FWIU we already have enough abandoned mines in the world to do all of our energy storage needs?
Here's a prompt for this one, for the AI:
"Let's think step-by-step: how much depth or volume of empty mines are needed to solve for US energy storage demand with gravitational potential energy storage drones on or off tracks in abandoned mines? Please respond with constants and code as Python SymPy code with a pytest test_main and Hypothesis @given decorator tests"
Show HN: Frogmouth – A Markdown browser for the terminal
Hi HN,
Frogmouth is a TUI to display Markdown files. It does a passable job of displaying Markdown, with code blocks and tables. No image support as yet.
It's very browser like, with navigation stack, history, and bookmarks. Works with both the mouse and keyboard.
There are shortcuts for viewing README.md files and other Markdown on GitHub and GitLab.
License is MIT.
Let me know what you think...
Could Frogmouth display Markdown cells in Jupyter Notebooks on the CLI?
FWIW, IPython is built with Python Prompt Toolkit and Jupyter_console is IPython for non-python kernels; `conda install -y jupyter_console; jupyter kernelspec -h`. But those only do `%logstart -o example.out.py`; Markdown in notebooks on the CLI is basically unheard-of.
I would like to run VSCode with a TUI like this in a terminal via SSH.
Does anyone know a VSCode extension to do this?
FWIU coc.vim works with vim and nvim and works with VSCode LSP support: https://github.com/neoclide/coc.nvim
There are many LSP implementations: https://langserver.org/
awesome-neovim Ctrl-F "DAP": https://github.com/rockerBOO/awesome-neovim#lsp
mason.nvim: https://github.com/williamboman/mason.nvim :
> Portable package manager for Neovim that runs everywhere Neovim runs. Easily install and manage LSP servers, DAP servers, linters, and formatters.
conda-forge
DAP: Debug Adapter Protocol https://microsoft.github.io/debug-adapter-protocol/implement...
COC has been mostly deprecated for native LSPs also. I run ~same pyright I'd use in VSCode in my neovim.
Stanford, Harvard data science no more
Do Stanford or Harvard have a UC BIDS: UC Berkeley Institute of Data Science?
> This is the [open] textbook for the Foundations of Data Science class at UC Berkeley: "Computational and Inferential Thinking: The Foundations of Data Science" http://inferentialthinking.com/ (JupyterBook w/ notebooks and MyST Markdown)
> [#1 Undergrad Data Science program, #2 ranked Graduate Statistics program]
Data literacy: https://en.wikipedia.org/wiki/Data_literacy :
> Data literacy is distinguished from statistical literacy since it involves understanding what data means, including the ability to read graphs and charts as well as draw conclusions from data.[6] Statistical literacy, on the other hand, refers to the "ability to read and interpret summary statistics in everyday media" such as graphs, tables, statements, surveys, and studies. [6]
Data Literacy and Statistical Literacy are essential for good leadership. For citizens to be capable of Evidence-Based Policy, we need Data Driven Journalism (DDJ) and curricular data science in the public high schools.
Debugging a Mixed Python and C Language Stack
GDB can decode Python stack traces as Python code, as long as you have have python-gdb.py from the Python source distribution in the same place as your Python executable.
From https://github.com/jupyterlab/debugger/issues/284 en ENH: Mixed Python/C debugging (GDB,) #284
> "Users can do [mixed mode] debugging with GDB (and/or other debuggers) and log the session in a notebook; in particular in order to teach.
> Your question is specifically about IDEs with support for mixed-mode debugging (with gdb), so I went looking for an answer:
> https://wiki.python.org/moin/DebuggingWithGdb (which is not responsive and almost unreadable on a mobile device) links to https://fedoraproject.org/wiki/Features/EasierPythonDebuggin... , which mentions the py-list, py-up and py-down, py-bt, py-print, and py-locals GDB commands that are also described in * https://devguide.python.org/gdb/ *
> https://wiki.python.org/moin/PythonDebuggingTools Ctrl-F "gdb" mentions: DDD, pyclewn (vim), trepan3k (which is gdb-like and supports breaking at c-line and also handles bytecode disassembly)
> Apparently, GHIDRA does not have a debugger but there is a plugin for following along with gdb in ghidra called https://github.com/Comsecuris/gdbghidra [...] https://github.com/Comsecuris/gdbghidra/blob/master/data/gdb... (zero dependencies)
> https://reverseengineering.stackexchange.com/questions/1392/... lists a number of GUIs for GDB; including voltronnn:
>> There's Voltron, which is an extensible Python debugger UI that supports LLDB, GDB, VDB, and WinDbg/CDB (via PyKD) and runs on macOS, Linux and Windows. For the first three it supports x86, x86_64, and arm with even arm64 support for lldb while adding even powerpc support for gdb. https://github.com/snare/voltron
> "The GDB Python API" https://developers.redhat.com/blog/2017/11/10/gdb-python-api... describes the GDB Python API.
> https://pythonextensionpatterns.readthedocs.io/en/latest/deb... may be helpful [for writing-a-c-function-to-call-any-python-unit-test]
> The GDB Python API docs: https://sourceware.org/gdb/onlinedocs/gdb/Python-API.html
> The devguide gdb page may be the place to list IDEs with support for mixed-mode debugging of Python and C/C++/Cython specifically with gdb?
These days, we have CFFI and Apache Arrow for C+Python etc.
https://wiki.python.org/moin/DebuggingWithGdb lists the GDB debugging symbol packages for various distros
FWIU Fedora GDB now optionally automatically installs debug syms; from attempting to debug TuxMath SDL with GDB before just installing the Flatpak.
https://developers.redhat.com/articles/2021/09/08/debugging-... (2021) describes more recent efforts to improve Python debugging with c extensions
Teach at a Community College
> Learning to teach
To learn, teach.
For contrast, there's also:
Those who can, do. Those who can't, teach.
Not that I 100% agree with that.
If none teach, few can do or teach.
"This is where teachers are paid the most" (2021) https://www.weforum.org/agenda/2021/10/teachers-pay-countrie... :
> Globally, teachers’ salaries vary hugely, according to the OECD’s Education At A Glance 2021 report
Twitter drops “Government-funded”/“state-affiliated“ from NPR, BBC, RT, Xinhua
Perhaps it would be more appropriate to retain the existing label, link to the resource indicating how much state financial support and/or editorial control, and also label other publishers by their known revenue methods:
- advertising supported
- directly politically supported
- PAC-funded
- SPAC-funded
- nonprofit organization [in territories x, y, z]
And maybe also their methods of journalism; in the interest of evidence-based policy by way of evidence-based journalism:
- Source and Method
- Motive and Intent
- Admissibility vs Publishability
- What else?
Media literacy > Media literacy education https://en.wikipedia.org/wiki/Media_literacy#Media_literacy_...
- Do they require disclosure of editorial conflicts of interest?
- Do they require disclosure of journalistic conflicts of interest?
- Do they allow citations?
- Do they require citations?
- Do they list the DOI string in the citation or a dx.doi.org URL for NewsArticles about ScholarlyArticles?
- Which contributors are paid?
- Which parent company owns the media outlet?
Physicists discover that gravity can create light
All quantum fields have their vacuum energy fluctuations. When we change the parameters of the space (e.g. stretch or compress it), the current state is no longer the ground state but becomes a so-called squeezed vacuum state.
This effect is used in laser physics to "split photons" in spontaneous parametric downconversion. That is, an intense laser changes the refractive index of a medium periodically. These oscillations generate a squeezed vacuum state.
Is there a corollary SQG Superfluid Quantum Gravity fluidic description of squeezed coherent states?
And what of a Particle-Wave-Fluid triality?
Noting that I attempted to link to relevant research and cite the source for fluidic corollaries but was prevented from contributing. https://westurner.github.io/hnlog/#comment-35661155
FWIW, so inspired, I continued to write a few better prompts for all of this.
"""Explain how specific (for example Fedi and Turok's) theories of gravitons and superfluidity differ from General Relativity, the Standard Model, prevailing theories of Quantum Gravity and dark matter and/or dark energy non-uniform correction coefficients, classical fluid dynamics, and quantum chaos theory in regards to specific phenomena in the quantum foam.
Also, are is there one wave function or are there many; and are they related by operators expressible with qubit quantum computers or are qudits and qutrits necessary to sufficiently model this domain of n body gravity gravitons in a fluid field?
If graviton fields result in photons, what are the conservation symmetry relations in regards to the exchange of gravitons for photons?"""
And then (though apparently currently one must remove "must use `dask_ml.model_selection.GridSearchCV`" presumably due to current response length limits of Google Bard):
"""Design an experiment as a series of steps and Python code to test (1) whether Bernoulli's equations describe gravitons in a superfluidic field; and also (2) there is conservational symmetry in exchange of gravitons and photons. The Python code must use `dask_ml.model_selection.GridSearchCV`, must use SymPy, define constants in the `__dict__` attribute of a class, return experimental output as a `dict`, have pytest tests and Hypothesis `@given` decorator tests, and a `main(argv=sys.argv)` function with a pytest `test_main`, and use `asyncio`."""
Manipulative Consent Requests
"fraudulent redefinition of consent"
Unconscionability: https://en.wikipedia.org/wiki/Unconscionability :
> Unconscionability (sometimes known as unconscionable dealing/conduct in Australia) is a doctrine in contract law that describes terms that are so extremely unjust, or overwhelmingly one-sided in favor of the party who has the superior bargaining power, that they are contrary to good conscience. Typically, an unconscionable contract is held to be unenforceable because no reasonable or informed person would otherwise agree to it. The perpetrator of the conduct is not allowed to benefit, because the consideration offered is lacking, or is so obviously inadequate, that to enforce the contract would be unfair to the party seeking to escape the contract.
Statute of frauds: https://en.wikipedia.org/wiki/Statute_of_frauds :
> A statute of frauds is a form of statute requiring that certain kinds of contracts be memorialized in writing, signed by the party against whom they are to be enforced, with sufficient content to evidence the contract. [1][2] […]
> Raising the defense: A defendant in a contract case who wants to use the statute of frauds as a defense must raise it as an affirmative defense in a timely manner. [7] The burden of proving that a written contract exists comes into play only when a statute of frauds defense is raised by the defendant.
Statute of frauds > United States > Uniform Commercial Code: https://en.m.wikipedia.org/wiki/Statute_of_frauds#Uniform_Co... :
> Uniform Commercial Code: In addition to general statutes of frauds, under Article 2 of the Uniform Commercial Code (UCC), every state except Louisiana has adopted an additional statute of frauds that relates to the sale of goods. Pursuant to the UCC, contracts for the sale of goods where the price equals $500 or more fall under the statute of frauds, with the exceptions for professional merchants performing their normal business transactions, and for any custom-made items designed for one specific buyer. [42]
IIUC, that means that if the USD amount of a contract for future performance is over $500 the court would regard a statute of frauds argument as just cause for dismissal? Or just goods?
Proliferation of AI weapons among non-state actors could be impossible to stop
Artificial intelligence arms race > Proposals for international regulation: https://en.wikipedia.org/wiki/Artificial_intelligence_arms_r...
Counterproliferation: https://en.wikipedia.org/wiki/Counterproliferation :
> In contrast to nonproliferation, which focuses on diplomatic, legal, and administrative measures to dissuade and impede the acquisition of such weapons, counterproliferation focuses on intelligence, law enforcement, and sometimes military action to prevent their acquisition. [2]
...
def hasRightTo_(person) -> bool:
def hasRight(thing, right) -> bool:
def haveEqualRights(persons:Iterable, rights:Iterable[Callable]) -> bool:
Practical methods of evidentiary proofs: Source, Method; Motive, Intent
Only one pair of distinct positive integers satisfy the equation m^n = n^m
I think taking logs is an unnecessary indirection in the given proof. If n^m = m^n then raising both sides to the power of 1/(nm) gives us n^(1/n) = m^(1/m). So we are looking for two distinct positive integers at which the function x ↦ x^(1/x) takes the same value. The rest of the proof then goes as before.
Differentiating the above function yields (1/x^2)(1-log(x))x^(1/x), which is positive when log(x) < 1 and negative when log(x) > 1. So the function has a maximum at e and decreases on either side of it. Therefore one of our integers must be less than e, and the other greater than it. For the smaller integer there are only two possibilities, 1 and 2. Using 1 doesn't give a solution since the equation x^(1/x) = 1 only has the solution 1. So the only remaining possibility for the smaller number is 2, which does yield the solution 2^4 = 4^2. Since x^(1/x) is strictly decreasing when x > e, there can't be any other solutions with the same value.
Logs will appear no matter what, at some point in the line of reasoning. They are only held back until differentiating in this approach. I think the expression for the derivative of logn/n was much nicer to grapple with.
Seeing as we're in the integer world, I'm not sure logs need to show up ever, there's probably a proof path around unique prime factorisations. E.g. it's more or less immediately obvious that n and m would need to share the same prime factors in the same ratios.
This back-and-forth makes me wonder, is it possible to write proofs about proofs? e.g., you cannot prove this result without logarithms or there must exist a proof that involves prime factors?
I suppose at least some proofs vaguely like that are possible because Gödel's incompleteness theorem is one, although I suspect that that same theorem puts some constraints on these "metaproofs."
Axiomatic system > Axiomatic method https://en.wikipedia.org/wiki/Axiomatic_system#Axiomatic_met...
Inference https://en.wikipedia.org/wiki/Inference ; inductive, deductive, abductive
Propositional calculus > Proofs in propositional calculus: https://en.wikipedia.org/wiki/Propositional_calculus
Quantum logic: https://en.wikipedia.org/wiki/Quantum_logic
https://twitter.com/westurner/status/1609495237738496000 :
> [ Is quantum logic the correct or a sufficient logic for propositional logic? ]
What are quantum "expectation values"; and how is that Axiomatic wave operator system different from standard propositional calculus?
Show HN: IPython-GPT, a Jupyter/IPython Interface to Chat GPT
This would be great in conjunction with e.g. papermill for running the same prompts over time and with the model as a parameter.
Do IPython-GPT or jetmlgpt work in JupyterLite in WASM in a browser tab?
I have only tested jetmlgpt using Jupyter notebooks.
JupyterLite How to docs: https://jupyterlite.readthedocs.io/en/latest/howto/index.htm...
"Create a custom kernel" https://jupyterlite.readthedocs.io/en/latest/howto/extension... :
> We recommend checking out how to create a server extension first
It may also be possible to just pip install the plain python package with micropip, or include it in a `jupyterlite build`; From https://github.com/jupyterlite/jupyterlite/issues/237#issuec... re: 'micropip':
%pip install $@
# __import__('piplite').install($@)
There's also micromamba, for mamba-forge packages (which are build with empack (emscripten) into WASM)Congrats on the launch!
I created Mito, a spreadsheet the lives inside of Jupyter. We've seen a ton of our users using Jupyter and ChatGPT playground side by side.
We're also experimenting with bringing chatgpt into Jupyter through the Mito spreadsheet. (https://blog.trymito.io/5-lessons-learned-from-adding-chatgp...)
Looking forward to trying this out.
Cocalc also has new "generate a Jupyter Notebook from a [ChatGPT] prompt" functionality. https://twitter.com/cocalc_com/status/1644427430223028225
'2023-04-19: LaTeX + ChatGPT "Help me fix this..." buttons into our LaTeX editor' https://cocalc.com/news/latex-chatgpt-help-me-fix-this-butto...
Building a ChatGPT-enhanced Python REPL
"Show HN: IPython-GPT, a Jupyter/IPython Interface to Chat GPT" https://news.ycombinator.com/item?id=35580959 lists a few REPLs
When you buy a book, you can loan it to anyone – a judge says libraries can’t
The headline isn't mentioned in the opinion piece? What the judge ruled was:
> An ebook recast from a print book is a paradigmatic example of a derivative work.
In other words, if a library buy a book, they can't just digitise it and share it without permission.
I'd go so far as to say the article directly contradicts the headline.
The author spends an enormous amount of time explaining how the rights of "end users" have been cemented, such as the right to change formats. If a consumer lends a book, they are, by definition, no longer the "end user" and none of that precedent applies.
It's probably true that anyone can lend a book but that's not an "end user" right. In fact, that allowance is the very reason libraries can exist in the first place.
The case and article are about something completely different (making copies, then distributing the copies). Odd headline.
> that allowance
Right to property: https://en.wikipedia.org/wiki/Right_to_property
Brain images just got 64 million times sharper
9.4 T is quite a strong magnetic field, as it is half the strength of the magnetic field used for plasma confinement in Tokamak Energy’s Demo4 facility. Scaling this technology for human-sized animals will likely be incredibly expensive, but the 5 micron accuracy is surely worth the investment.
7T is already regularly used for human research, and approval for human usage has been granted for 10.5T and I believe for 11.7T (though I'm not sure how many images they've gotten out of that yet).
Yes it is incredibly expensive, but it is in fact already done.
https://thedebrief.org/impossible-photonic-breakthrough-scie... :
> For decades, that [Abbe diffraction] limit has operated as a sort of roadblock to engineering materials, drugs, or other objects at scales smaller than the wavelength of light manipulating them. But now, the researchers from Southampton, together with scientists from the universities of Dortmund and Regensburg in Germany, have successfully demonstrated that a beam of light can not only be confined to a spot that is 50 times smaller than its own wavelength but also “in a first of its kind” the spot can be moved by minuscule amounts at the point where the light is confined.
> According to that research, the key to confining light below the previous impermeable Abbe diffraction limit was accomplished by “storing a part of the electromagnetic energy in the kinetic energy of electric charges.” This clever adaptation, the researchers wrote, “opened the door to a number of groundbreaking real-world applications, which has contributed to the great success of the field of nanophotonics.”
> “Looking to the future, in principle, it could lead to the manipulation of micro and nanometre-sized objects, including biological particles,” De Liberato says, “or perhaps the sizeable enhancement of the sensitivity resolution of microscopic sensors.”
"Electrons turn piece of wire into laser-like light source" (2022) https://news.ycombinator.com/item?id=33493885
Could such inexpensive coherent laser light sources reduce medical and neuroimaging costs?
"A simple technique to overcome self-focusing, filamentation, supercontinuum generation, aberrations, depth dependence and waveguide interface roughness using fs laser processing" https://scholar.google.com/scholar?start=10&hl=en&as_sdt=5,4... :
> Several detrimental effects limit the use of ultrafast lasers in multi-photon processing and the direct manufacture of integrated photonics devices, not least, dispersion, aberrations, depth dependence, undesirable ablation at a surface, limited depth of writing, nonlinear optical effects such as supercontinuum generation and filamentation due to Kerr self-focusing. We show that all these effects can be significantly reduced if not eliminated using two coherent, ultrafast laser-beams through a single lens - which we call the Dual-Beam technique. Simulations and experimental measurements at the focus are used to understand how the Dual-Beam technique can mitigate these problems. The high peak laser intensity is only formed at the aberration-free tightly localised focal spot, simultaneously, suppressing unwanted nonlinear side effects for any intensity or processing depth. Therefore, we believe this simple and innovative technique makes the fs laser capable of much more at even higher intensities than previously possible, allowing applications in multi-photon processing, bio-medical imaging, laser surgery of cells, tissue and in ophthalmology, along with laser writing of waveguides.
TL Transfer Learning might be useful for training a model to predict e.g. [portable] low-field MRI with NIRS Infrared and/or Ultrasound? FWIU, "Mind2Mind" is one way to ~train a GAN from another already-trained GAN?
From https://twitter.com/westurner/status/1609498590367420416 :
> Idea: Do sensor fusion with all available sensors timecoded with landmarks, and then predict the expensive MRI/CT from low cost sensors
> Are there implied molecular structures that can be inferred from low-cost {NIRS, Light field, [...]} sensor data?
> Task: Learn a function f() such that f(lowcost_sensor_data) -> expensive_sensor_data
FWIU OpenWater has moved to NIRS+Ultrasound for ~ live in surgery MRI-level imaging and now treatment?
FWIU certain Infrared light wavelengths cause neuronal growth; and Blue and Green inhibit neuronal growth.
What are the comparative advantage and disadvantages of these competing medical imaging and neuroimaging capabilities?
Can a technology like this at some point be used for brain uploads? Or like a cheaper and better version of cryonics - to "backup" the brain until it can be resurrected in some way?
If quantum information is never destroyed – and classical information is quantum information without the complex term i – perhaps our brain states are already preserved in the universe; like reflections in water droplets in the quantum foam.
Lagrangian points, non-intersecting paths through accretion discs, and microscopic black holes all preserve data - modulated energy; information - for some time before reversible or unreversible transformation.
Perhaps Superfluid quantum gravity can afford insight into the interior topology of black holes and other quantum foam phenomena?
(Edit)
Coping: https://en.wikipedia.org/wiki/Coping
Defence mechanisms § Level 4: mature: https://en.wikipedia.org/wiki/Defence_mechanism#Level_4:_mat...
MRI brain images become 64M times sharper
Other thread from earlier today: "Brain images just got 64 million times sharper" https://news.ycombinator.com/item?id=35615778
There’s no universal cordless power tool battery – why?
From "USB-C is about to go from 100W to 240W, enough to power beefier laptops" https://news.ycombinator.com/item?id=27295621 :
> What are the costs to add a USB PD module to an electronic device? https://hackaday.com/2021/04/21/easy-usb%E2%80%91c-power-for...
> - [ ] Create an industry standard interface for charging and using [power tool,] battery packs; and adapters
From "Zero energy ready homes are coming" https://news.ycombinator.com/item?id=35065366 :
# USB power specs (DC)
7.5w = 1.5amp * 5volts # USB
15w = 3a * 5v # USB-C
100w = 5a * 20v # USB-C PD
240w = 5a * 48v # USB-C PD 3.1
Others could also send emails to power tool manufacturers about USB-C PD support and evolving the spec to support power tool battery packs.Aura – Python source code auditing and static analysis on a large scale (2022)
Unfortunately, there hasn't been a release since 2021, Similarly no commits to the branches master, or dev in the past ~14 months.
Hello, author of Aura here. The project is in fact active! But in a different branch called "ambience". Which is a very big refactor into transforming Aura (which is now designed to be run locally as tui) into server/web application. It would allow to automatically monitor and audit all used python packages in an organization by using an http reverse proxy to intercept python package installations. It's taking me currently long time to finish that big refactor as I am currently the only active developer there so apologies if the project seems to be abandoned, I'm just hesitating to merge the changes from ambience branch into main (which is what people see) as the new refactor is not stable yet as compared to master & dev as that was tested and tuned on the whole PyPI.
Very early alpha version is available here: https://ambience.sourcecode.ai if someone is interested in checking it out.
Could aura scan packages at the pulp pypi proxy? https://github.com/pulp
I haven't used pulp so I am not sure but yes in theory. Several schemes are currently supported via URIs (pypi://, git://, http(s):// etc...) so if the destination to scan can be formatted as one of the already supported URI schemes then you can already scan it. URI providers are also using plugin architecture so adding a new one for better integration with pulp (such as autodiscovering packages) should be trivial. Thank you for the suggestion, the pulp project looks interesting and I would definitely check it out!
Gitea can also (scan and build and test and) host python packages [1], conda packages [2], container images, etc.
[1] https://docs.gitea.io/en-us/usage/packages/pypi/
[2] https://docs.gitea.io/en-us/usage/packages/conda/
https:// URLs probably already solve for scanning Python packages hosted by Gitea and/or Pulp with Aura.
From https://news.ycombinator.com/item?id=33563857 :
> Additional lists of static analysis, dynamic analysis, SAST, DAST, and other source code analysis tools: https://news.ycombinator.com/item?id=24511280 https://analysis-tools.dev/tools?languages=python
A new approach to computation reimagines artificial intelligence
Before reading the paper's obliquely mentioned in the article: How is this different from embeddings?
After reading the papers: How is this different from embeddings?
All three of the papers linked to are closed access. (Did I miss one?)
If anyone can link a favorite review paper on this approach, you'd be doing us a favor.
Edit: one of them does link to a preprint version: https://arxiv.org/abs/2203.04571
Great, and the implementation (Neuro-vector-symbolic architectures):
"A Neuro-vector-symbolic Architecture for Solving Raven's Progressive Matrices" (2023) https://arxiv.org/abs/2203.04571 :
> […] our proposed neuro-vector-symbolic architecture (NVSA) [implements] powerful operators on high-dimensional distributed representations that serve as a common language between neural networks and symbolic AI. The efficacy of NVSA is demonstrated by solving the Raven's progressive matrices datasets. Compared to state-of-the-art deep neural network and neuro-symbolic approaches, end-to-end training of NVSA achieves a new record of 87.7% average accuracy in RAVEN, and 88.1% in I-RAVEN datasets. Moreover, compared to the symbolic reasoning within the neuro-symbolic approaches, the probabilistic reasoning of NVSA with less expensive operations on the distributed representations is two orders of magnitude faster.
A number system invented by Inuit schoolchildren
I dislike these pseudo-scientific claims about alternative number systems and methods of paper and pencil arithmetic:
> Because of the tally-inspired design, arithmetic using the Kaktovik numerals is strikingly visual. Addition, subtraction and even long division become almost geometric. The Hindu-Arabic digits are an awkward system, Bartley says, but “the students found, with their numerals, they could solve problems a better way, a faster way.”
I think the students can be praised for having come up with simple to understand and write number system that corresponds to the conventions for counting in Alaskan Inuit language, and it seems appropriate to capture these notations in upcoming Unicode standards.
However, spending time learning base 20 arithmetic has obvious disadvantages that the article ignores. The times tables, memorized in grade school and fundamental to paper and pencil calculations, are now four times larger. Base 20 is not a popular notation for numbers. One important advantage of the number system (Hindu-Arabic) that most of the world uses is that most of the world uses it. I grew up with inches and degrees Fahrenheit and had to learn the metric system to pursue my science education. I'm glad I didn't have to learn how to count as well. We shouldn't make it harder for these kids to enjoy the rest of the world's books, journals, and internet resources about math and science.
> “the students found, with their numerals, they could solve problems a better way, a faster way”
> Base 20 is not a popular notation for numbers […] We shouldn't make it harder for these kids
So much of what you object to is that something they’ve found more intuitive and engaging isn’t what unintuitive disengaging stuff they’ll encounter. But developing intuition for math is far more valuable than developing conformance to how it’s supposed to be done. Who cares if that intuition is developed with some idiosyncrasy from what you consider normal? The math is math, the principles are consistent, the knowledge is transferable. Insisting they learn the same things a different way is totally arbitrary and counterproductive.
I think parent's point is not cultural as you are implying but rather practical. Learning in a base that nobody uses could be easier today and an hindrance later.
I think, all in all, this should not be a big deal. For the gifted kid, they'll find a way to adapt and become the next Einstein. As for the ungifted, it might give them a better leg up and allow them to perform better than they would have, so it's probably a plus anyways.
Finger binary: https://en.wikipedia.org/wiki/Finger_binary :
> Finger binary is a system for counting and displaying binary numbers on the fingers of either or both hands. Each finger represents one binary digit or bit. This allows counting from zero to 31 using the fingers of one hand, or 1023 using both: that is, up to 2**5−1 or 2**10−1 respectively.
- "How to count to 1000 on two hands" by 3blue1brown https://youtu.be/1SMmc9gQmHQ
- "Polynesian People Used Binary Numbers 600 Years Ago - Scientific American" https://www.scientificamerican.com/article/polynesian-people...
What is the comparative value of radixes like Binary, Octal, andHexadecimal compared to Decimal (radix 10)?
Perhaps a radix like eπI would be more useful; though some amost-mystic physicists do tend to radix 9: "nonary" (which is actually ~ also radix-3).
List of numeral systems > By culture / time period, By type of notation https://en.wikipedia.org/wiki/List_of_numeral_systems :
> Numeral systems are classified here as to whether they use positional notation (also known as place-value notation), and further categorized by radix or base.
Digital computing hardware design
I always make a point of demonstrating finger-binary with the number 132. 4 will suffice, though.
California DMV wants to issue car titles as NFTs
Can this also be solved with W3C Verified Claims and ~ https://blockcerts.org/ (and an append-only, multiply distributed/synchronized/replicated/redundantly-backed-up permissioned database/Blockchain)?
And then, again, what is the TPS of the DB? How many read and write Transactions Per Second do the candidate datastores support with or without strong Consistency, Availability, and/or Partition Tolerance? And then what about cryptographic Data Integrity assurances.
FWIU, issuing shares on a Blockchain is challenging due to lost keys / shared cryptographic keys and reissuance?
"Possible" and "beneficial" are separate questions entirely.
The whole point of government is to be a central authority. That's the DMV's entire job: to be a central authority for driver's licenses and vehicle titles. Their database isn't open for everybody to modify, and it also shouldn't be. It also shouldn't be open for everybody to read, because we have a right to privacy.
What benefit does the DMV get from hosting its records on a blockchain? And what happens if that blockchain stops attracting miners? Does the DMV have to mine it? Does the price of mining rigs start to affect how much it costs to register your car? What happens if they get hit by a 51% attack?
Your average Californian is definitely not a crypto user, and doesn't have a digital wallet. One in twenty don't even speak English. What happens when somebody loses their private key? If you have to keep those keys in a conventional database, then a blockchain makes zero sense except as a money-waster.
There is nothing wrong with them just running a regular database, with normal SRE stuff like backups and replication. The California DMV has a lot of problems. But "how do we store records?" isn't one of them.
It takes patience to keep politely explaining the inadequacies of "5 people and the manager and the owner share the sole root backup credentials - which aren't cryptographic keys, so there is no way to tell whether data has been rewritten without the correct key - and tape backup responsibilities; and we have no way to verify the online database against the tape backups [so, can we write it off now]"
Notes on centralized databases with Merkel hashes : https://westurner.github.io/hnlog/ Ctrl-F "trillian", "iavl"
If you've ever set up SQL replication before, you wouldn't be claiming it's adequate.
Are RDS and CloudSQL good enough? What does Accumulo do differently?
Show HN: Skip the SSO Tax, access your user data with OSS
As the former CTO of an Insurtech and Fintech startup I always had the “pleasure” to keep regulators and auditors happy. Think of documenting who has access to what, quarterly access reviews, yearly audits and so on…
Like many others we couldn’t justify the Enterprise-plan for every SaaS tool to simply get access to SSO and SCIM/SAML APIs. For Notion alone the cost would have nearly doubled to $14 per user per month. That’s insane! Mostly unknown to people, SSO Tax also limits access to APIs that are used for managing user access (SCIM/SAML).
This has proven to be an incredibly annoying roadblock that prevented me from doing anything useful with our user data: - You want to download the current list of users and their permissions? Forget about it! - You want to centrally assign user roles and permissions? Good luck with that! - You want to delete user accounts immediately? Yeah right, like that's ever gonna happen!
It literally cost me hours to update our access matrix at the end of every quarter for our access reviews and manually assigning user accounts and permissions.
I figured, there must be a better way than praying to the SaaS gods to miraculously make the SSO Tax disappear (and open up SCIM/SAML along the way). That’s why I sat down a few weeks ago and started building OpenOwl (https://github.com/AccessOwl/open_owl). It allows me to just plug in my user credentials and automatically download user lists, including permissions from SaaS tools.
Granted, OpenOwl is still a work in progress, and it's not perfect. At the moment it's limited to non-SSO login flows and covers only 7 SaaS vendors. My favorite part is that you can configure integrations as “recipes”. The goal was for anybody to be able to add new integrations (IT managers and developers alike). Therefore you ideally don’t even have to write any new code, just tell OpenOwl how the new SaaS vendor works.
What do you think? Have you dealt with manually maintaining a list of users and their permissions? Could this approach get us closer to overcoming parts of the SSO Tax?
For those wondering what the "SSO Tax" is, it refers to the excessive pricing practiced by SaaS providers to access the SSO feature on their product.
A documented rant has made the rounds at https://sso.tax , which lists all vendors and their pricing of SSO.
I always thought this was insane, but now I wonder if "SSO Pricing"/tax is just the "real price" and the "Base pricing" is really the new free trial? Of course the SSO/real pricing is too high, and everyone negotiates it down, but the point is I suspect the "base pricing" is just a trial teaser that's probably not sustainable for many vendors in terms of margins. I'm just guessing here, maybe someone with some inside insight from one of these vendors can advise.
I'd rather consider it "SME pricing" vs "Enterprise pricing". Typically only companies above a certain size use SSO systems, and even larger ones require it for everything. Coincidentally bigger companies are also willing to pay more, so putting a high price on SSO enables SaaS to profit from those deep pockets without pricing themselves out of the market for smaller companies.
Schools, colleges, and universities typically have SSO but no budget or purchase authority.
Just slap an education discount on it and call it a day. There are plenty of reasons to do that anyway, you want students to get trained on your software and use it in their formative years as much as possible.
Many go even further and just give the product away for free for educational institutions and individual students (Github, Jetbrains and Tableau come to mind as examples)
https://github.com/doncicuto/glim :
> Glim is a simple identity access management system that speaks some LDAP and has a REST API to manage users and groups
"Proxy LDAP to limit scope of access #60" https://github.com/doncicuto/glim/issues/60
Twitter Is Blocking Likes and Retweets that Mention Substack
In December they tried to ban mentioning any non-Twitter usernames:
https://techcrunch.com/2022/12/18/twitter-wont-let-you-post-...
Isn't that completely contradictory to their supporting free speech mission?
It sure looks like someone has hijacked Elon and Jack and run off with those operations.
Maybe you have to add the year to the search query to find search results from before when Donny got kicked out fairly for serial disrespect.
Become a more effective censorship apparatus.
What "supporting free speech mission" are you referring to?
https://investor.twitterinc.com/contact/faq/default.aspx :
> What is Twitter's mission statement?
> The mission we serve as Twitter, Inc. is to give everyone the power to create and share ideas and information instantly without barriers. Our business and revenue will always follow that mission in ways that improve – and do not detract from – a free and global conversation
"Defending and respecting the rights of people using our service" https://help.twitter.com/en/rules-and-policies/defending-and...
Ha nice. That doesn't actually say "free speech" but point taken. Of course mission statements are always meaningless, but still, point taken.
PEP 684 was accepted – Per-interpreter GIL in Python 3.12
This is great and was a huge effort by Eric Snow and collaborators to make it happen.
To be clear, this is not "PEP 703 – Making the Global Interpreter Lock Optional in CPython", by Sam Gross. PEP 703 is more ambitious and add quite a bit of complexity to the CPython implementation. PEP 684 mostly cleans things up and encapsulates global state better. The most distruptive change is to allow immortal objects. I hope PEP 703 can go in too, even with the downsides (complexity, bit slower single threaded performance).
"PEP 703 – Making the Global Interpreter Lock Optional in CPython" (2023) https://github.com/python/peps/blob/main/pep-0703.rst https://peps.python.org/pep-0703/
colesbury/nogil https://github.com/colesbury/nogil :
docker run -it nogil/python
docker run -it nogil/python-cuda
Detection of Common Cold from Speech Signals Using Deep Neural Network
"Detection of COVID-19 from voice, cough and breathing patterns: Dataset and preliminary results" (2021) https://scholar.google.com/scholar?hl=en&as_sdt=5,47&sciodt=...
Pandas 2.0
I'm curious if there will be any appreciable performance gains here that are worthwhile. FWIW, last I checked[0], Polars still smokes Pandas in basically every way.
I’ve moved entirely to Polars (which is essentially Pandas written in Rust with new design decisions) with DuckDB as my SQL query engine. Since both are backed by Arrow, there is zero copy and performance on large datasets is super fast (due to vectorization, not just parallelization)
I keep Pandas around for quick plots and legacy code. I will always be grateful for Pandas because there truly was no good dataframe library during its time. It has enabled an entire generation of data scientists to do what they do and built a foundation — a foundation which Polars and DuckDB are now building on and have surpassed.
How well does Polars play with many of the other standard data tools in Python (scikit learn, etc.)? Do the performance gains carry over?
It works as well as Pandas (realize that scikit actually doesn’t support Pandas — you have to cast a dataframe into a Numpy array first)
I generally work with with Polars and DuckDB until the final step, when I cast it into a data structure I need (Pandas dataframe, Parquet etc)
All the expensive intermediate operations are taken care of in Polars and DuckDB.
Also a Polars dataframe — although it has different semantics — behaves like a Pandas dataframe for the most part. I haven’t had much trouble moving between it and Pandas.
You do not have to cast pandas DataFrames when using scikit-learn, for many years already. Additional in recent version there has been increasing support for also returning DataFrames, at least with transformers and checking column names/order.
Yes that support is still not complete. When you pass a Pandas dataframe into Scikit you are implicitly doing df.values which loses all the dataframe metadata.
There is a library called sklearn-pandas which doesn’t seem to be mainstream and dev has stopped since 2022.
From pandas-dataclasses #166 "ENH: pyarrow and optionally pydantic" https://github.com/astropenguin/pandas-dataclasses/issues/16... :
> What should be the API for working with pandas, pyarrow, and dataclasses and/or pydantic?
> Pandas 2.0 supports pyarrow for so many things now, and pydantic does data validation with a drop-in dataclasses.dataclass replacement at pydantic.dataclasses.dataclass.
Model output may or may not converge given the enumeration ordering of Categorical CSVW columns, for example; so consistent round-trip (Linked Data) schema tool support would be essential.
CuML is scikit-learn API compatible and can use Dask for distributed and/or multi-GPU workloads. CuML is built on CuDF and CuPY; CuPy is a replacement for NumPy arrays on GPUs with 100x relative performance.
CuPy: https://github.com/cupy/cupy :
> CuPy is a NumPy/SciPy-compatible array library for GPU-accelerated computing with Python. CuPy acts as a drop-in replacement to run existing NumPy/SciPy code on NVIDIA CUDA or AMD ROCm platforms.
> CuPy is an open-source array library for GPU-accelerated computing with Python. CuPy utilizes CUDA Toolkit libraries including cuBLAS, cuRAND, cuSOLVER, cuSPARSE, cuFFT, cuDNN and NCCL to make full use of the GPU architecture.
> The figure shows CuPy speedup over NumPy. Most operations perform well on a GPU using CuPy out of the box. CuPy speeds up some operations more than 100X. Read the original benchmark article Single-GPU CuPy Speedups on the RAPIDS AI Medium blog
CuDF: https://github.com/rapidsai/cudf
CuML: https://github.com/rapidsai/cuml :
> cuML is a suite of libraries that implement machine learning algorithms and mathematical primitives functions that share compatible APIs with other RAPIDS projects.*
> cuML enables data scientists, researchers, and software engineers to run traditional tabular ML tasks on GPUs without going into the details of CUDA programming. In most cases, cuML's Python API matches the API from scikit-learn.
> For large datasets, these GPU-based implementations can complete 10-50x faster than their CPU equivalents. For details on performance, see the cuML Benchmarks Notebook.
FWICS there's now a ROCm version of CuPy, so it says CUDA (NVIDIA only) but also compiles for AMD. IDK whether there are plans to support Intel OneAPI, too.
What of the non-Arrow parts of other pandas-compatible and not pandas-compatible DataFrame libraries can be ported back to Pandas (and R)?
RFdiffusion: Diffusion model generates protein backbones
So I guess this means easier drug discovery? Honesty those wiggly diagrams are meaningless to me I have no bio background
Easier drug discovery is what they tell public and grant agencies. In a roundabout way it’s true. Maybe. Many other hurdles still exist. But what this and other similar tools really are, is significantly advancing basic science in creating our own protein designs.
Before alphafold changed this field, creating your own protein design was considered an insane task (not impossible, bakers lab and others have done it a couple times). But these tools (now we have multiple) allow you to create new proteins From scratch that can do exactly what you want (caveats galore). New enzymes that can catalyze reactions never found in nature for example.
Before this all we could do was take proteins that already exist in nature and modify them. So you can imagine how new this world is.
Large Language models can also generate novel and working protein structures that adhere to a specified purpose https://www.nature.com/articles/s41587-022-01618-2
Can optical tweezers construct such proteins; or is there a more efficient way?
Optical tweezers: https://en.wikipedia.org/wiki/Optical_tweezers
"'Impossible' photonic breakthrough: scientist manipulate light at subwavelength scale" https://thedebrief.org/impossible-photonic-breakthrough-scie... :
> have successfully demonstrated that a beam of light can not only be confined to a spot that is 50 times smaller than its own wavelength but also “in a first of its kind” the spot can be moved by minuscule amounts at the point where the light is confined.
> According to that research, the key to confining light below the previous impermeable Abbe diffraction limit was accomplished by “storing a part of the electromagnetic energy in the kinetic energy of electric charges.” This clever adaptation, the researchers wrote, “opened the door to a number of groundbreaking real-world applications, which has contributed to the great success of the field of nanophotonics.”
> “Looking to the future, in principle, it could lead to the manipulation of micro and nanometre-sized objects, including biological particles,” De Liberato says, “or perhaps the sizeable enhancement of the sensitivity resolution of microscopic sensors.”
"Digging into DNA Repair with Optical Tweezer Technology" https://www.genengnews.com/topics/digging-into-dna-repair-wi...
It's much easier than that! Living cells already have ribosomes that construct proteins and all the other molecular machinery needed to go from DNA sequence to assembled protein. You can order a DNA sequence online and put it into e-coli or yeast cells and those cells will make that protein for you.
TIL about mail-order CRISPR kits. "Mail-Order CRISPR Kits Allow Absolutely Anyone to Hack DNA" (2017) https://www.scientificamerican.com/article/mail-order-crispr...
Protein production: https://en.wikipedia.org/wiki/Protein_production
Tissue Nanotransfection reprograms e.g. fibroblasts into neurons and endothelial cells (for ischemia) using electric charge. Are there different proteins then expressed? Which are the really useful targets?
> The delivered cargo then transforms the affected cells into a desired cell type without first transforming them to stem cells. TNT is a novel technique and has been used on mice models to successfully transfect fibroblasts into neuron-like cells along with rescue of ischemia in mice models with induced vasculature and perfusion
> [...] This chip is then connected to an electrical source capable of delivering an electrical field to drive the factors from the reservoir into the nanochannels, and onto the contacted tissue
https://en.wikipedia.org/wiki/Tissue_nanotransfection#Techni...
Are there lab safety standards for handling yeast or worse? https://en.wikipedia.org/wiki/Gene_drive
"Bacterial ‘Nanosyringe’ Could Deliver Gene Therapy to Human Cells" (2023) https://www.scientificamerican.com/article/bacterial-nanosyr... :
> In a paper published today in Nature, researchers report refashioning Photorhabdus’s syringe—called a contractile injection system—so that it can attach to human cells and inject large proteins into them. The work could provide a way to deliver various therapeutic proteins into any type of cell, including proteins that can “edit” the cell’s DNA. “It’s a very interesting approach,” says Mark Kay, a gene therapy researcher at Stanford University who was not involved in the study. “Where I think it could be very useful is when you want to express proteins that can do genome editing” to correct or knock out a gene that is mutated in a genetic disorder, he says.
> The nano injector could provide a critical tool for scientists interested in tweaking genes. “Delivery is probably the biggest unsolved problem for gene editing,” says study investigator Feng Zhang, a molecular biologist at the McGovern Institute for Brain Research at the Massachusetts Institute of Technology and the Broad Institute of M.I.T. and Harvard. Zhang is known for his work developing the gene editing system CRISPR-Cas9. Existing technology can insert the editing machinery “into a few tissues, blood and liver and the eye, but we don’t have a good way to get to anywhere else,” such as the brain, heart, lung or kidney, Zhang says. The syringe technology also holds promise for treating cancer because it can be engineered to attach to receptors on certain cancer cells.
From "New neural network architecture inspired by neural system of a worm" (2023) https://news.ycombinator.com/item?id=34715188 :
> "I’m skeptical that biological systems will ever serve as a basis for ML nets in practice"
>> First of all, ML engineers need to stop being so brainphiliacs, caring only about the 'neural networks' of the brain or brain-like systems. Lacrymaria olor has more intelligence, in terms of adapting to exploring/exploiting a given environment, than all our artificial neural networks combined and it has no neurons because it is merely a single-cell organism [1].
Which proteins code for organisms that compute?
Llama.cpp 30B runs with only 6GB of RAM now
Author here. For additional context, please read https://github.com/ggerganov/llama.cpp/discussions/638#discu... The loading time performance has been a huge win for usability, and folks have been having the most wonderful reactions after using this change. But we don't have a compelling enough theory yet to explain the RAM usage miracle. So please don't get too excited just yet! Yes things are getting more awesome, but like all things in science a small amount of healthy skepticism is warranted.
Didn't expect to see two titans today: ggerganov AND jart. Can ya'll slow down you make us mortals look bad :')
Seeing such clever use of mmap makes me dread to imagine how much Python spaghetti probably tanks OpenAI's and other "big ML" shops' infra when they should've trusted in zero copy solutions.
Perhaps SWE is dead after all, but LLMs didn't kill it...
You can mmap from python.
The CPython mmap module docs: https://docs.python.org/3/library/mmap.html
zero_buffer (CFFI, 2013) https://github.com/alex/zero_buffer/blob/master/zero_buffer....
"Buffers on the edge: Python and Rust" (2022) https://alexgaynor.net/2022/oct/23/buffers-on-the-edge/ :
> If you have a Python object and want to obtain its buffer, you can do so with memoryview in Python or PyObject_GetBuffer in C. If you’re defining a class and want to expose a buffer, you can do so in Python by… actually you can’t, only classes implemented in C can implement the buffer protocol. To implement the buffer protocol in C, you provide the bf_getbuffer and bf_releasebuffer functions which are called to obtain a buffer from an object and when that buffer is being released, respectively.
iocursor (CPython C API, ~Rust std::io::Cursor) https://github.com/althonos/iocursor
Arrow Python (C++) > On disk and MemoryMappedFile s: https://arrow.apache.org/docs/python/memory.html#on-disk-and...
"Apache Arrow: Read DataFrame With Zero Memory" (2020) https://towardsdatascience.com/apache-arrow-read-dataframe-w...
pyarrow.Tensor: https://arrow.apache.org/docs/python/generated/pyarrow.Tenso...
ONNX is built on protocolbuffers/protobufs (google/protobufs), while Arrow is built on google/flatbuffers.
FlatBuffers https://en.wikipedia.org/wiki/FlatBuffers :
> It supports “zero-copy” deserialization, so that accessing the serialized data does not require first copying it into a separate part of memory. This makes accessing data in these formats much faster than data in formats requiring more extensive processing, such as JSON, CSV, and in many cases Protocol Buffers. Compared to other serialization formats however, the handling of FlatBuffers requires usually more code, and some operations are not possible (like some mutation operations).
We updated our RSA SSH host key
The fact that this key was apparently not stored in an HSM, and that GH employees had access to this private key (allowing them to accidentally push it) means that effectively all communication with GH since the founding of the company has to be considered compromised. This basically means that, depending on your level of paranoia, you will have to review all code that has ever interacted with Github repositories, and any code pushed or pulled from private repositories can no longer be considered private.
Github's customers trust GH/MS with their code, which, for most businesses, is a high value asset. It wouldn't surprise me if this all results in a massive lawsuit. Not just against GH as a company, but also to those involved (like the CISO). Also, how on earth was it never discovered during an audit that the SSH private key was plain text and accessible? How has GH ever been able to become ISO certified last year [0], when they didn't even place their keys in a HSM?
Obviously, as a paying customer I am quite angry with GH right now. So I might be overreacting when I write this, but IMO the responsible persons (not talking about the poor dev that pushed the key, but the managers, CISO, auditors, etc.) should be fined, and/or lose their license.
[0] https://github.blog/2022-05-16-github-achieves-iso-iec-27001...
SSH uses ephemeral keys. It's not enough to have the private key and listen to the bytes on the wire, you have to actively MITM the connection. A github employee who has access to the private key and enough network admin privileges to MITM your connection already has access to the disk platters your data is saved on.
Regarding the secrecy of the data you host at github, you should operate under the assumption that a sizeable number of github employees will have access to it. You should assume that it's all sitting unencrypted on several different disk platters replicated at several different geographically separated locations. Because it is.
One of the things that you give up when you host your private data on the cloud is controlling who and under what circumstances people can view or modify your private data. If the content of the data is enough to sink you/your company without remedy you should not store it on the cloud.
I would also add that your ability to pretend to be the client to the server is also limited, if ssh key based client authentication is used. This means that even if the host key is leaked, an attacker will not be able to push in the name of the attacked client. The attacker will be able to pretend to be the server to the client, and thus be able to get the pushed code from the client (even if the client just added one patch, the attacker can pretend to be an empty repo server side and receive the entire repo.
If ssh token based auth is used, it's different of course, because then the server gets access to the token. Ideally Github would invalidate all their auth tokens as well.
The fun fact is that a token compromise (or any other attack) can still happen any point in the future with devices that still have outdated ssh keys. That's a bit unfortunate as no revocation mechanism exists for ssh keys... ideally clients would blacklist the ssh key, given the huge importance of github.
WKD also lacks key revocation and CT Certificate Transparency.
E.g. keybase could do X.509 like Certificate Transparency.
:
$ keybase pgp -h
NAME:
keybase pgp - Manage keybase PGP keys
USAGE:
keybase pgp <command> [arguments...]
COMMANDS:
gen Generate a new PGP key and write to local secret keychain
pull Download the latest PGP keys for people you follow.
update Update your public PGP keys on keybase with those exported from the local GPG keyring
select Select a key from GnuPG as your own and register the public half with Keybase
sign PGP sign a document.
encrypt PGP encrypt messages or files for keybase users
decrypt PGP decrypt messages or files for keybase users
verify PGP verify message or file signatures for keybase users
export Export a PGP key from keybase
import Import a PGP key into keybase
drop Drop Keybase's use of a PGP key
list List the active PGP keys in your account.
purge Purge all PGP keys from Keybase keyring
push-private Export PGP keys from GnuPG keychain, and write them to KBFS.
pull-private Export PGP from KBFS and write them to the GnuPG keychain
help, h Shows a list of commands or help for one command
$ keybase pgp drop -h
NAME:
keybase pgp drop - Drop Keybase's use of a PGP key
USAGE:
keybase pgp drop <key-id>
DESCRIPTION:
"keybase pgp drop" signs a statement saying the given PGP
key should no longer be associated with this account. It will **not** sign a PGP-style
revocation cert for this key; you'll have to do that on your own.
/?q=PGP-style revocation cert
https://www.google.com/search?q=PGP-style+revocation+cert :- "Revoked a PGP key, is there any way to get a revocation certificate now?" https://github.com/keybase/keybase-issues/issues/2963
- "Overview of Certification Systems: X.509, CA, PGP and SKIP"
...
- k8s docker vault secrets [owasp, inurl:awesome] https://www.google.com/search?q=k8s+docker+vault+secrets+owa... https://github.com/gites/awesome-vault-tools
- Why secrets shouldn't be passed in $ENVIRONMENT variables; though e.g. the "12 Factor App" pattern advises to parametrize applications mostly with environment variables that show in /proc/pid/environ but not /proc/pid/cmdline
W3C DID supports GPG proofs and revocation IIRC:
"9.6 Key and Signature Expiration" https://www.w3.org/TR/did-core/#key-and-signature-expiration
"9.8 Verification Method Revocation" https://www.w3.org/TR/did-core/#verification-method-revocati...
Blockerts is built upon W3C DID and W3C Verified Credentials, W3C Linked Data Signatures, and Merkel trees (and JSON-LD). From the Blockerts FAQ https://www.blockcerts.org/guide/faq.html :
> How are certificates revoked?
> Even though certificates can be issued to a cohort of people, the issuer can still revoke from a single recipient. The Blockcerts standard supports a range of revocation techniques. Currently, the primary factor influencing the choice of revocation technique is the particular schema used.
> The Open Badges specification allows a HTTP URI revocation list. Each id field in the revokedAssertions array should match the assertion.id field in the certificate to revoke.
Re: CT and W3C VC Verifiable Credentials (and DNS record types for cert/pubkey hashes that must also be revoked; DoH/DoT + DNSSEC; EDNS): https://news.ycombinator.com/item?id=32753994 https://westurner.github.io/hnlog/#comment-32753994
"Verifiable Credential Data Integrity 1.0: Securing the Integrity of Verifiable Credential Data" (Working Draft March 2023) > Security Considerations https://www.w3.org/TR/vc-data-integrity/#security-considerat...
If a system does not have key revocation it cannot be sufficiently secured.
Ask HN: Where can I find a primer on how computers boot?
As a developer, I recently encountered challenges with GRUB and discovered I lacked knowledge about my computer's boot process. I realized terms like EFI partition, MBR, GRUB, and Bootloader were unfamiliar to me and many of my colleagues. I'm seeking introductory and easy-to-understand resources to learn about these concepts. Any recommendations would be appreciated!
Booting process of Linux: https://en.wikipedia.org/wiki/Booting_process_of_Linux
Booting process of Windows NT since Vista: https://en.wikipedia.org/wiki/Booting_process_of_Windows_NT_...
UEFI > Secure Booting, Boot Stages: https://en.wikipedia.org/wiki/UEFI#Boot_stages
The EFI system partition is conventionally /boot/efi on a Linux system; and there's a signed "shim loader" that GRUB launches, which JMP- launches the kernel+initrd after loading the initrd into RAM (a "RAM drive") and mounting it as the initial root filesystem /, which is pivot_root'd away from after the copy of /sbin/init (systemd) mounts the actual root fs and launches all the services according to the Systemd unit files in order according to a topological sort given their dependency edges: https://en.wikipedia.org/wiki/EFI_system_partition
Runlevels: https://en.wikipedia.org/wiki/Runlevel
runlevel 5 is runlevel 3 (multi-user with networking) + GUI. On a gnome system, GDM is the GUI process that is launched. GDM launches the user's Gnome session upon successful login. `systemctl restart gdm` restarts the GDM Gnome Display Manager "greeter" login screen, which runs basically runs ~startx after `bash --login`. Systemd maps the numbered runlevels to groups of unit files to launch:
telinit 6 # reboot
telinit 3 # kill -15 GDM and all logged in *GUI* sessions
You can pass a runlevel number as a kernel parameter by editing the GRUB menu item by pressing 'e' if there's not a GRUB password set; just the number '3' will cause the machine to skip starting the login greeter (which may be what's necessary to troubleshoot GPU issues). The word 'rescue' as a kernel parameter launches single-user mode, and may be what is necessary to rescue a system failing to boot. You may be able to `telinit 5` from the rescue runlevel, or it may be best to reboot.>The EFI system partition is conventionally /boot/efi on a Linux system;
Still often mounted as /boot/efi but physically residing on the root of the hidden ESP volume as EFI/boot & EFI/ubuntu, for distros like ubuntu.
So when mounted you have something like /boot/efi/boot/bootx64.efi and boot/efi/ubuntu/grub etc. if visible and accessible.
Internet Control Message Protocol (ICMP) Remote Code Execution Vulnerability
Show HN: PicoVGA Library – VGA/TV Display on Raspberry Pi Pico
TV/VGA + Serial Console on an RP2040 would be cool.
- rs232
- UART
- USB-TTL w/ configurable (3.3v/5v) voltage levels
- (optional) VGA output
- (optional) WiFi/Bluetooth (RP2040W Pi Pico W)
The OpenWRT hardware/port.serial wiki page explains how to open a serial console with screen and a tty /dev; and about 3.3V/5V bricking a router and/or the cable: https://openwrt.org/docs/techref/hardware/port.serial#use_yo...
PiKVM may have console support but with a Pi 3/4+, not a $4/$6 RP2040 /W with or without low-level WiFi+Bluetooth
"Raspberry Pi Pico Serial Communication Example (MicroPython)" https://electrocredible.com/raspberry-pi-pico-serial-uart-mi...
> The RP2040 microcontroller in Raspberry Pi Pico has two UART peripherals, UART0 and UART1. The UART in RP2040 has the following features: [pinout]
The reason for this port is that I'm using PicoVGA as a VGA terminal for an 8bit retro computer. I'm using USB host mode on the Pico for keyboard and mouse input. I found a Pico W for sale for a reasonable price + shipping so now I can add wifi support.
Can it run a ~GNUscreen text console attached to a serial port attached to pins on the other side of the Pico, and maybe a BLE keyboard/controller? (And how, while I'm at it; Thonny has an AST parse tree menu item)
I don't see why not. Pico has a few dedicated UART pins. Bluetooth support is pretty new, though.
2023-02-10; bluekitchen/btstack: https://github.com/raspberrypi/pico-sdk/commit/c8ccefb9726b8...
> - RS232
RS232: https://en.wikipedia.org/wiki/RS-232
Serial Port: https://en.wikipedia.org/wiki/Serial_port
> - UART
UART: Universal Asynchronous Receiver Transmitter: https://en.wikipedia.org/wiki/Universal_asynchronous_receive...
> - USB-TTL w/ configurable (3.3v/5v) voltage levels
USB-to-serial adapter: https://en.wikipedia.org/wiki/USB-to-serial_adapter
> Most commonly the USB data signals are converted to either RS-232, RS-485, RS-422, or TTL-level UART serial data.
- SunFounder Thales Pi Pico Kit > Components > Slide Switch, Resistor, : https://docs.sunfounder.com/projects/thales-kit/en/latest/co... Kepler has the 2040W: https://docs.sunfounder.com/projects/kepler-kit/en/latest/
/? how to limit voltage to 3.3v: https://www.google.com/search?q=how+to+limit+voltage+to+3.3v
- Zener diode > Voltage shifter: https://en.wikipedia.org/wiki/Zener_diode#Voltage_shifter
- Voltage regulator > DC voltage stabilizer: https://en.wikipedia.org/wiki/Voltage_regulator#DC_voltage_s...
- TIL electronics.stackexchange has CircuitLab built-in: https://electronics.stackexchange.com/questions/241537/3-3v-... . Autodesk TinkerCAD is a neat, free, web-based circuit simulator, too; though it supports Arduino and bbc:micro but not (yet?) Pi Pico like Wokwi.
There is at least one RP2040 JS simulator:
Wokwi appears to support {MicroPython, CircuitPython,} on {Pi Pico} and also Rust on ESP32:
Wokwi > New Pi Pico project: https://wokwi.com/projects/new/pi-pico
Wokwi > New Pi Pico + MicroPython project: https://wokwi.com/projects/new/micropython-pi-pico
wokwi/rp2040js : https://github.com/wokwi/rp2040js
https://www.hackster.io/Hack-star-Arduino/raspberry-pi-pico-...
> - (optional) WiFi/Bluetooth (RP2040W Pi Pico W)
/? "pi pico" bluetooth 5.2 [BLE] https://www.google.com/search?q=%22pi+pico%22+bluetooth+5.2 :
- From https://www.cnx-software.com/2023/02/11/raspberry-pi-pico-w-... :
>> The Raspberry Pi Pico W board was launched with a WiFi 4 and Bluetooth 5.2 module based on the Infineon CYW43439 wireless chip in June 2022
bluekitchen/btstack is the basis for the Bluetooth support i Pi Pico SDK 1.5.0+: https://github.com/bluekitchen/btstack
> MicroPython
awesome-micropython: https://github.com/mcauser/awesome-micropython
https://github.com/pfalcon/awesome-micropython :
- There is a web-based MicroPython REPL, and it only works over HTTP due to WebSockets WSS, so it's best to clone that file locally: http://micropython.org/webrepl/ https://github.com/micropython/webrepl/
awesome-circuitpython: https://github.com/adafruit/awesome-circuitpython
- microsoft/vscode-python-devicesimulator lacks maintainers and Pi Pico Support: https://github.com/microsoft/vscode-python-devicesimulator/w...
https://vscode.dev/ is the official hosted copy of VSCode as WASM in a browser tab. YMMV with which extensions work after compilation to WASM and without WASI (node/fs,) in a browser tab.
From "MicroPython officially becomes part of the Arduino ecosystem" (2022) https://news.ycombinator.com/item?id=33699666 :
> [in addition to the Arduino IDE support for Pi Pico now] For VSCode, there are a number of extensions for CircuitPython and MicroPython:
> joedevivo.vscode-circuitpython: https://marketplace.visualstudio.com/items?itemName=joedeviv...
> Pymakr https://github.com/pycom/pymakr-vsc/blob/next/GET_STARTED.md
> Pico-Go: https://github.com/cpwood/Pico-Go
https://scratch.mit.edu/ has bbc:microbit (and LEGO Boost) extensions when you click plus at the lower left; but not (yet?) Pi Pico support: https://github.com/LLK https://github.com/LLK/scratch-gui/commit/421d673e714a367ff2... https://github.com/microbit-more/mbit-more-v2
FDIC – SVB FAQ
> The FDIC will pay uninsured depositors an advance dividend within the next week.
Depending on the amount, this can possibly make things a lot better for the startups banking there.
Where does the money come from? I mean that non-sarcastically, confess I've not figured out the way the banking system in the large works.
edit : thanks all for coherent clear responses.
A lot of SVB's assets can be sold immediately in robust markets (e.g. treasuries)
If that was true the bank wouldn't be insolvent.
And yet: https://dfpi.ca.gov/wp-content/uploads/sites/337/2023/03/DFP...
If you have $160B in deposits but only $150B in liquid assets, you're still insolvent. Doesn't mean the assets aren't liquid.
You are only insolvent if you have $160B in withdrawals. If the deposits sat until the bonds matured, the $150B in currently liquid assets would be worth more than $150B.
How would the strict separation between savings deposits and investment banking from Glass-Steagull (1933) (which was repealed by GLBA in 1999) banking regulations have prevented this?
From "1999 Repeal of Glass-Steagall was the worst deregulation enacted in US history" (2022) https://news.ycombinator.com/item?id=30206570 :
> Yeah what was the deal with that dotcom correction in the early 2000s? Did banks invest differently after GLBA said that they can gamble against peoples' savings deposits (because they created a 'sociallist' $100b credit line, called it FDIC, and things like that don't happen anymore)
> Decline of the Glass-Steagall Act: https://en.wikipedia.org/wiki/Decline_of_the_Glass%E2%80%93S...
> Dot-com bubble: https://en.wikipedia.org/wiki/Dot-com_bubble
An Update on USDC and Silicon Valley Bank
23% of the reserve, or $9.7B is held in cash.
I think on Thursday, SVB saw cash outflow requests over $40b in one day. USDC peg is so far below 1 right now there will definitely be a lot of par arbitrage Monday trying to withdraw.
Hopefully it doesn’t take longer than a day for them to settle treasury trades, and maybe some people have lost the keys to their crypto. They could also possibly have a line of credit with instant settlement assuming they’re truly solvent with marketable securities at market prices.
3 month (or less) Treasury Bills are damn close to cash. I don't believe they lose value in practice, not even in the rapidly raising interest rate environment of the last year.
The problem at Silvergate and Silicon Valley Banks is that they bought 10 year or 30 year bonds, not 3 month bills. Its a completely different composition, with completely different properties.
Why lock up the money for so long in such an unstable time as this? Better interest rates?
> Better interest rates?
You betcha.
Here's the rates for March 2021. 3M would have gotten you 0.05% APY, while 10Y gets you 1.45% APY.
https://home.treasury.gov/resource-center/data-chart-center/...
Yeah, it sounded like a perfect plan if no customers wanted to get their funds within the next 10 years. ¯\_(ツ)_/¯
No that's not it. You can sell the bonds any time. But now they're worth considerably less because interest rates are much higher now. They locked themselves in at rate that are now terrible for 10 years.
So weird to think that near-zero interest rates are the best you can do for the next 10 years. I'm pretty sure everyone way saying "lock in this low interest rate before they go up again!"
On Treasuries over what windowed series? https://fred.stlouisfed.org/tags/series?t=treasury%3Byield+c...
Is this the one?
"10-Year Treasury Constant Maturity Minus 3-Month Treasury Constant Maturity" https://fred.stlouisfed.org/series/T10Y3M
Urgent: Sign the petition now
Trump & republicans cut banking regulations and increased capital thresholds from 50 to 250 billion for banks to comply with an FDIC stress test as mandated by Dodd-Frank (which was enacted after the mortgage crisis). SVB would have not gone bankrupt had they been required to pass this stresstest. This change was widely supported by tech companies and banks and here we are again with a bank engaging in risky behavior holding out their hand for taxpayers to bail them out: NO.
Stress test (financial) > Bank stress test , Payment and settlement systems stress test https://en.wikipedia.org/wiki/Stress_test_(financial)#Bank_s...
List of bank stress tests > Americas https://en.wikipedia.org/wiki/List_of_bank_stress_tests#Amer...
https://www.google.com/search?q=increased+capital+thresholds...
From "Transparency & Accountability - EGRRCPA (S. 2155) Rulemakings" https://www.fdic.gov/transparency/egrrcpa.html :
> The FDIC is responsible for a number of rulemakings under the Economic Growth, Regulatory Relief, and Consumer Protection Act (EGRRCPA). This page provides links to proposed and final rules and related documents.
"FDIC Releases Economic Scenarios for 2022 Stress Testing" (2022) https://www.fdic.gov/news/press-releases/2022/pr22019.html
How can the scenarios and policies be improved?
129-year-old vessel still tethered to lifeboat found on floor of Lake Huron
I'd love to dive there someday.
The NOAA page on the wreck (https://thunderbay.noaa.gov/shipwrecks/ironton.html) just has depth listed as "TBA", but the Ohio that it collided with is in >100m.
assuming it's in similar range, that's serious depth. Way into hypoxic trimix zone (ie: you'll be breathing gas that'd kill you if you were breathing it on the surface)
Looks fascinating, but almost certainly out of reach for all but the most extreme technical divers.
Crazy how 100m is such a short distance in most walks of life except for underwater.
It’s seriously non intuitive to a layman the whole pressure thing.
https://en.wikipedia.org/wiki/Decompression_sickness#Ascent_... :
> DCS [Decompression Sickness] is best known as a diving disorder that affects divers having breathed gas that is at a higher pressure than the surface pressure, owing to the pressure of the surrounding water. The risk of DCS increases when diving for extended periods or at greater depth, without ascending gradually and making the decompression stops needed to slowly reduce the excess pressure of inert gases dissolved in the body.
DCS > Prevention > Underwater diving: https://en.wikipedia.org/wiki/Decompression_sickness#Underwa... :
> Decompression time can be significantly shortened by breathing mixtures containing much less inert gas during the decompression phase of the dive (or pure oxygen at stops in 6 metres (20 ft) of water or less). The reason is that the inert gas outgases at a rate proportional to the difference between the partial pressure of inert gas in the diver's body and its partial pressure in the breathing gas; whereas the likelihood of bubble formation depends on the difference between the inert gas partial pressure in the diver's body and the ambient pressure. Reduction in decompression requirements can also be gained by breathing a nitrox mix during the dive, since less nitrogen will be taken into the body than during the same dive done on air. [85]
Google Groups has been left to die
Google Groups has survived 22 years. How many self-hosted Discourse forums can say the same? If you're working on a new, small side-project like a new formal evaluator, are you going to want to spend your limited administrative bandwidth onboarding new contributors, responding to feature suggestions, talking to users, etc, or are you going to want to spend it trying to wrangle self-hosted forum administration and figure out email delivery? the article has a one-line tossed off aside at the end to say "Certainly self-hosted FOSS communities can die, but these are functions of community activity itself rather than the service they’re hosted on", but without considering that the absolute most crucial time for community resources to survive is through periods of lackluster or nonexistent community involvement. If you have a Google Group full of, say, formal methods programming experts, then it can spend 5 years fallow and still be chock-full of absolutely vital historical resources and even spring back into life as people start using it again for discussion. That sort of longevity just can't exist if someone stopped paying the hosting bills for their self-hosted TLA discourse forum 3 years in.
AtariAge.com[1] - first created in 1988, or 25 years ago. Run by some bloke on the internet.
The problem for Google Groups is that no-one can perf-farm it any more, which is the sole reason for projects to survive at Google...
perf-farm ? I think I might know what you mean, but could you unpack?
Not the person who used the term, but I kind of wish I had been. It's excellent. At FAANG-ish companies where "impact" is the key to promotion, there's an inevitable tendency for engineers to prefer working on things that are highly visible like starting new projects or adding new features, vs. things that are highly useful like stability, testing, or code quality. Most internally promoted E6s or above got that way by initiating multiple projects, then dumping them on others when they've served their perf-review-enhancing purpose. "Perf farming" is as good a term as I've seen for it.
> Most internally promoted E6s or above got that way by initiating multiple projects, then dumping them on others
Well, Jeff and Sanjay got to DE/Fellows this way, so this is a good role model. Unfortunately doesn’t scale.
Zero energy ready homes are coming
Contrary to what I am seeing in comments on HN, this is not just for single family detached housing. If you go to the ZERH website, under Program Requirements it says:
Versions 1 and 2, Single Family, Multifamily, and Manufactured Homes
https://www.energy.gov/eere/buildings/zero-energy-ready-home...
Additionally, one of the photos in the featured article has the following caption:
Lopez Community Land Trust built this 561-square-foot affordable home on Lopez Island, Washington, to the high-performance criteria of DOE's Zero Energy Ready Home program that delivers a $20-per-month average monthly energy bill.
I've said for years we need to go back to more passive solar design as our baseline standard because it is both more energy efficient and more comfortable for humans. It protects the environment while improving quality of life.
It's a myth that you can't see gains in both areas.
I'm also happy to see the caption because we need to do a better job of supporting high quality of life in small towns, not just big cities. Trends in recent decades mean small towns suck because they lack services and big cities suck if you aren't one of the wealthy.
That wasn't always true and programs like this can improve and strengthen both small town and urban environments while mitigating the burden suburbs currently represent.
Passive solar building design : https://en.wikipedia.org/wiki/Passive_solar_building_design#...
Low-energy house: https://en.wikipedia.org/wiki/Low-energy_house
List of low-energy building techniques: https://en.wikipedia.org/wiki/List_of_low-energy_building_te...
/? Passive solar home (tbm=isch image search) https://www.google.com/search?q=passive+solar+home&tbm=isch
/? Passive solar house: https://youtube.com/results?search_query=passive+solar+house
/? passive solar house: https://www.pinterest.com/search/pins/?q=passive%20solar%20h...
(Edit)
Maximum solar energy is on the equatorial side of the house.
Full-sun plants prefer maximum solar energy.
10ft (3m) underground it's about 75°F (24°C) all year. Geothermal systems leverage this. (Passive) Walipini greenhouses are partially or fully underground or in a hillside, but must also manage groundwater seepage and flooding; e.g. with a DC solar sump pump and/or drainage channels filled with rock.
Passive solar greenhouses (especially in China and now Canada) have a natural or mounded earthen wall thermal mass on one side, and they lower wool blankets over the upside-down wing airfoil -like transparent side at night and when it's too warm (with a ~4HP motor).
TIL an aquarium heater can heat a tank of water working as a thermal mass in a geodesic growing dome; which can be partially-buried or half-walled with preformed hempcrete block.
Round structures are typically more resilient to wind:
Shear stress > Beam shear: https://en.wikipedia.org/wiki/Shear_stress
Deformation (physics) > Strain > Shear strain: https://en.wikipedia.org/wiki/Deformation_(physics)#Shear_st...
GH topic: finite-element-analysis https://github.com/topics/finite-element-analysis
GH topic: structural-analysis: https://github.com/topics/structural-engineering https://github.com/topics/structural-analysis
What open source software is there for passive home design and zero-energy home design?
Round house: https://www.pinterest.com/search/pins/?q=round%20house
The shearing force due to wind on structures with corners (and passive rooflines) causes racking and compromise of structural integrity; round homes apparently fare best in hurricanes.
Walipini passive solar green houses: https://en.wikipedia.org/wiki/Walipini
Earthship passive solar homes: https://en.wikipedia.org/wiki/Earthship
Underground living: https://en.wikipedia.org/wiki/Underground_living
Root cellar passive refrigeration: https://en.wikipedia.org/wiki/Root_cellar
Ground source heat pump (Geothermal heat pump) https://en.wikipedia.org/wiki/Ground_source_heat_pump
Solar-assisted heat pump: https://en.wikipedia.org/wiki/Solar-assisted_heat_pump
(Geothermal Power = Geothermal Electricity) != (Geothermal Heating)
Geothermal power: https://en.wikipedia.org/wiki/Geothermal_power
Geothermal Heating is so old, https://en.wikipedia.org/wiki/Geothermal_heating
...
AC-to-DC (rectifier; GaNprime, GaN) and DC-to-AC (inverter) are inefficient conversions: it wastes electricity as heat.
Residential and Commercial AC electrical systems have a GFCI ground loop (for the ground pin on standard AC adapters)
WAV: Watts = Amps * Volts
# USB power specs (DC)
7.5w = 1.5amp * 5volts # USB
15w = 3a * 5v # USB-C
100w = 5a * 20v # USB-C PD
240w = 5a * 48v # USB-C PD 3.1
# 110v/120v AC: 15amp; Standard Residential AC in North America:
1500w = 15a * 100v # Microwave oven
1650w = 15a * 110v
1440w = 12a * 120v # AC EV charger
# 110/120 AC: 20amp
2400w = 20a * 120v # Level 1 EV charger
# 240v AC plug: Dryer, Oven, Stove, EV
4800w = 20a * 240v
7200w = 30a * 240v # Public charging station
9600w = 40a * 240v # Level 2 EV charget
14400w = 60a * 240v
120000w = 300a * 400v # Supercharger v2
120kW = 300a * 400v
1000000w = 1000a * 1000v # Megacharger
1000kW = 1000a * 1000v
1MW = 1000a * 1000v
USB > Power related standards: https://en.wikipedia.org/wiki/USB#Power-related_standardsCharging station > Charging time > charger specs table: https://en.wikipedia.org/wiki/Charging_station#Charging_time
(Trifuel) generators do not have catalytic converters.
Wood stoves must be sufficiently efficient; and can be made so with a catalytic combustor or a returning apparatus (and/or thermoelectrics to convert heat to electricity).
/? Catalytic combustor (wood stove) https://www.google.com/search?q=%22catalytic+combustor%22
Wood-burning stove > Safety and pollution considerations > US pollution control requirements: https://en.wikipedia.org/wiki/Wood-burning_stove#US_pollutio...
Gravitational potential energy is less lossy than CAES Compressed Air Energy Storage is less lossy than thermal salt is less lossy than chemical batteries.
'Pain mound' low tech compost driven water heaters are neat too
This is a pile of vaguely relevant facts...
I think in my HN comments I'd prefer a couple of facts and a conclusion, with citations for more detail.
Stochastic gradient descent written in SQL
Title here is wrong. Title in article and headings in article are right: ONLINE gradient descent
It's specifically not stochastic. From the article:
Online gradient descent
Finally, we have enough experience to implement online gradient descent. To keep things simple, we will use a very vanilla version:
- Constant learning rate, as opposed to a schedule.
- Single epoch, we only do one pass on the data.
- Not stochastic: the rows are not shuffled. ⇠ ⇠ ⇠ ⇠
- Squared loss, which is the standard loss for regression.
- No gradient clipping.
- No weight regularisation.
- No intercept term.
Hehe I was wondering if someone would catch that. Rest assured, I know the difference between online and stochastic gradient descent. I admit I used stochastic on Hacker News because I thought it would generate more engagement.
What are some adversarial cases for gradient descent, and/or what sort of e.g. DVC.org or W3C PROV provenance information should be tracked for a production ML workflow?
Gradient descent: https://en.wikipedia.org/wiki/Gradient_descent
Stochastic gradient descent: https://en.wikipedia.org/wiki/Stochastic_gradient_descent
Online machine learning: https://en.wikipedia.org/wiki/Online_machine_learning
adversarial gradient descent site:github.com inurl:awesome : https://www.google.com/search?q=awesome+adversarial+gradient...
https://github.com/EthicalML/awesome-production-machine-lear...
Robust machine learning: https://en.wikipedia.org/wiki/Robustness_(computer_science)#...
Robust gradient descent
We built model & data provenance into our open source ML library, though it's admittedly not the W3C PROV standard. There were a few gaps in it until we built an automated reproducibility system on top of it, but now it's pretty solid for all the algorithms we implement. Unfortunately some of the things we wrap (notably TensorFlow) aren't reproducible enough due to some unfixed bugs. There's an overview of the provenance system in this reprise of the JavaOne talk I gave here https://www.youtube.com/watch?v=GXOMjq2OS_c. The library is on GitHub - https://github.com/oracle/tribuo.
U.S. corn-based ethanol worse for the climate than gasoline, study finds
As others mentioned, this is not a new finding. There is a lot of academic and industrial/economic analysis on this subject. Ethanol from corn was never a good idea economically for anyone other than corn farmers. It is also bad environmentally and places additional competitive demand on our food supply (choice and competition between corn going into food or fuel supply).
Similar analysis has been done on other crop feed-stocks (switch grass, etc.) and none of the analysis has shown that crop derived ethanol is a good idea.
I'm not a leading expert on this subject by any means, but I am a PhD chemical engineer and did a lot of research into alternative fuels science and economics about ten years ago. I'm not up on the latest research and analysis but the fundamental thermodynamics and carbon cycles of crop-based ethanol have not changed much in the past ten years or so.
Corn farmers from Iowa. Where they hold the first caucus of the interminable election season.
No American politician or political party can afford to offend corn farmers. We should count ourselves lucky that our gasoline is only 10% corn.
Can Ethanol be sustainably produced from corn?
Which other crops have sufficient margin given commodity prices?
Can solar and goats and wind and IDK algae+co2 make up the difference?
Is solar laser weeding a more sustainable approach?
What rotations improve the compost and soil situation?
> Can solar and goats and wind and IDK algae+co2 make up the difference?
Can row covers etc be made from corn; from cellulose and what other sustainable inputs (ideally that are waste outputs)?
Compostable thermoformable biopolymers for food packaging
FWIU transparent wood "windows" are made out of treated cellulose? How do they compare with glass and plastic in terms of transparency for full spectrum UV for e.g. plant growth, sanitization, and vitamin D?
> Is solar laser weeding a more sustainable approach?
From "Solar-Powered Plant Protection Equipment: Perspective and Prospects" (2022) https://www.mdpi.com/1996-1073/15/19/7379 :
>> The unmanned system (robot) is well suited for weeding operations, and it helps to minimize the required workforce and herbicide usage while weeding. Two solar-powered weeders, EcoRobot and AVO robot models, are developed by Ecorobotix, Switzerland (Figure 2). These models work more effectively in row crops based on the detection of weeds (>85%), and a herbicides is applied precisely on the weeds to destroy them. The solar power used in EcoRobot and AVO models is 380 W and 1150 W, respectively, and they have a working time of 8 and 12 h once fully charged by solar panels [64].
> What rotations improve the compost and soil situation?
Crop rotation: https://en.wikipedia.org/wiki/Crop_rotation
SymPy makes math fun again
I like sagemath. They have integration with sympy, and they also have many features making python simpler when you want to do mathematical things.
There's latex display feature when in notebook. It looks good, can even have symbolical matrices and display them as you would've expected.
It's on par with Wolfram notebook, but you get to program in python instead of some propietary language.
If you got some really long formula, it's easier to see you got it correctly by displaying it symbolically in latex.
They have other things to smooth out the parts of python which are annoying in math, like exponentiation works with ^, and dividing integers returns a rational number instead of floating point or an integer.
> They have other things to smooth out the parts of python which are annoying in math, like exponentiation works with ^, and dividing integers returns a rational number instead of floating point or an integer.
SageMath also has a multivariate inequality solver IIRC. `*` is repeated multiplication (exponentiation) in Python. If you require preprocessing to translate ^ to *, you don't have valid Python code that'll run with any other interpreter.
Is it easier to import SageMath from a plain Python script now; with conda repackaging?
Type promotion in Python:
from rational import Rational
assert Rational(1, 4) / 2 == Rational(1, 8)
Python 2 had a __future__ import to do floatdiv instead of floordiv: from __future__ import division
assert 3 / 2 == 1.5
assert 3 // 2 == 1
assert 4 / 2 == 2.0
Python 3 returns floats from ints or decimals IIRC: assert 3 / 2 == 1.5
assert 3.0 / 2 == 1.5
assert 3 // 2 == 1
assert 4 / 2 == 2.0
> Is it easier to import SageMath from a plain Python script now; with conda repackaging?
Yes, it's easy. It was always possible, since fortunately none of the Sage library implementation uses the Sage preprocessor. You could always do "sage -python" and you get the Python interpreter that Sage uses, and can import Sage via "import sage.all". Somebody asked me exactly this question at a colloquium talk I gave on Sage yesterday, and here's my answer/demo:
https://youtu.be/dy792FEh1ns?t=1140
In the demo, I use a Jupyter kernel that runs the Python included in Sage-9.8.
Disclaimer: I created the Sage preparser long ago, in order to make Sage more palatable to mathematicians. In particular, it was inspired by audience feedback during a live demo I did at PyCon 2005.
> They have other things to smooth out the parts of python
I should add that this is an interesting chunk of Python code, and we've always planned to separate it out from Sage as a separate library than could be used with Sympy, etc.
https://github.com/sagemath/sage/blob/52a81cbd161ef4d5895325...
It does a surprising number of little clever things we realized would be a good idea over years, and it has some neat ideas from various Python PEP's like [1..10] that were rejected from Python, but are very useful when expressing mathematics. It's also annoying when you just want to format your code using prettier (say), and you can't because Sage isn't Python. :-(
Recently I've been working with the %%ipytest magic command for running pytest in notebooks.
Is there a %%sage IPython magic method that passes code through the preparse.py input transform?
Please tell us what features you’d like in news.ycombinator (2007)
Client-side encryption for Gmail in Google Workspace is now generally available
Not available to users with personal Google Accounts.
Why not?
I’m speculating about how this works, but it probably doesn’t make sense for general-purpose email.
For emails between members of one workspace/domain, you can conceptually perform key exchange to implement client-side encryption. You also don’t have spam concerns within members of a workspace.
There’s no general way to do key exchange between different email domains, and you do have spam concerns (which to address requires scanning the content).
There’s no protocol for doing such things over different email domains in a way that’s compatible and interoperable across software stacks. This technology is proprietary, and would make spam challenging to deal with (which again is not a problem within one workspace).
It’s also not clear to me whether this is really client-side encryption, so much as encryption at a layer higher than the email stack. (It’s pretty hard to do client-side encryption in the browser, in a way that matches user expectations. How do you store/retrieve the encryption key?)
OoenPGP.js is open source and developed by ProtonMail https://openpgpjs.org/ https://github.com/openpgpjs/openpgpjs
A number of Chrome (and I think also Firefox) extensions include their own local copy of OpenPGP.js for use with various webmail services, including GMail.
WKD (and HKP) depends upon HTTPS without cert pinning, FWIU: https://wiki.gnupg.org/WKD
How does an email client use WKD?
1. A user selects a recipient for an email.
2. The email client uses the domain part of the email address to construct which server to ask.
3. HTTPS is used to get the current public key.
The email client is ready to encrypt and send now.
An example:
https://intevation.de/.well-known/openpgpkey/hu/it5sewh54rxz33fwmr8u6dy4bbz8itz4 is the direct method URL for "bernhard.reiter@intevation.de
How/where exactly does it store the user's private keys?
The initial reaction that I have to client-side encryption with a web browser is: how does the client fetch/obtain the encryption key? (For signing outbound messages, or decrypting/verifying inbound.) Local storage wouldn't be appropriate as the sole storage for something like an asymmetric keypair, and it wouldn't provide the behavior that users expect: being able to log in and use a service from any web browser.
No one interacts with a web browser in such a way that it would be appropriate for the browser's local state to be the sole storage of something important like a private key. So this necessitates a service that the client interacts with during login to fetch its private key.
If you can log in with a browser, then that implies that you're fetching the key from somewhere. In the context of hosted Gmail, then that means that Gmail will be storing and providing your key. In which case it's nominally client-side encryption -- and the system that stores and provides your key might be totally separate from the email systems (and tightly controlled) -- but it's still not what I'd think of as client-side encryption generally. It's not really client-side encryption if the service that you're using to store data is the same service (from your perspective) as the one that stores/provides your key (those responsibilities might be separated from an implementation perspective, but they aren't from a user's perspective).
The service can employ various techniques like encrypting its copy of my private key using a symmetric key derived from my passphrase (and then discarded) -- and this decryption could perhaps be done client-side in the browser -- but ultimately the service still has the ability to obtain my key (at time of next login, if not any time).
If the service that provides a client-side encryption key could be divorced from a particular use-case -- e.g., if I could use my own pluggable key-providing service with Gmail, which my browser uses to fetch my key on login -- then the model would make a bit more sense. (You'd still have to trust a service like Gmail not to steal your key, but at least you don't have to trust it to store the key as well.)
Would it be possible to implement a standardized pluggable key server model? It seems plausible. Imagine that there was a standard HTTP protocol for this purpose, kind of like how WebAuthN is standardized. I log into Gmail, link it to my key server <https://keys.jcrites.example.com>, verify the connection, and then when logging into Gmail, then the web page is allowed to make HTTPS calls to <https://keys.jcrites.example.com> to either fetch my asymmetric key, or alternatively make calls to decrypt or encrypt specific content.
In the former case, it implements client-side encryption using a key server I control, while trusting Gmail not to misuse the key; in the latter case, it implements encryption using a keyserver that I control, and makes remote calls for each encryption operation. In that case Gmail would never have my key, although while browsing Gmail.com it could probably make arbitrary requests to the key server to perform operations using the key -- but you could log what operations are performed.
Looks like you are correct: WebUSB and WebBluetooth and WebAuth don't already cover HSM use cases?
/? secure enclave browser
But WebCrypto: "PROPOSAL: Add support for general (hardware backed) cryptographic signatures and key exchange #263" https://github.com/w3c/webcrypto/issues/263
Show HN: Classic FPS Wolfenstein 3D brought in the browser via Emscripten
TuxMath and TuxTyping are FOSS games written with SDL.
Giving your experience with porting Wolfenstein 3D SDL to emscripten, would it be easier to rewrite TuxMath given the exercise XML files or port it to WASM/emscripten (and emscripten-forge)?
Notably, the TuxMath RPM currently segfaults with recent Fedora but the Flatpak (which presumably statically-ships it's own copy of SDL) does work fine. https://flathub.org/apps/details/com.tux4kids.tuxmath https://github.com/tux4kids/tuxmath
(Other someday priorities: Looking at SensorCraft, wanting to port it to (JupyterLite WASM) notebooks w/ jupyter-book)
The latest Wolfenstein where you're girls that respawn only if tagteam is good, too; Wolfenstein: Youngblood. https://en.wikipedia.org/wiki/Wolfenstein
Porting SDL(2) games with Emscripten to get an actual WASM output is pretty straightforward. Getting the game to run properly is another story. If the game has additional dependencies, you have to build them separatly with Emscripten and link them in finally. Heavy use of the filesystem has to be commented out as well (browser tab is a sandbox).
emscripten-forge builds a ~condaenv of recipes with empack; like conda-forge: https://github.com/emscripten-forge/empack
There's not yet an SDL recipe yet: https://github.com/emscripten-forge/recipes/tree/main/recipe...
Re: emscripten fs implementations: https://github.com/emscripten-core/emscripten/issues/15041#i... https://github.com/jupyterlite/jupyterlite/issues/315
Show HN: Mathesar – open-source collaborative UI for Postgres databases
Hi HN! We just released the public alpha version of Mathesar (https://mathesar.org/, code: https://github.com/centerofci/mathesar).
Mathesar is an open source tool that provides a spreadsheet-like interface to a PostgreSQL database.
I was originally inspired by wanting to build something like Dabble DB. I was in awe of their user experience for working with relational data. There’s plenty of “relational spreadsheet” software out there, but I haven’t been able to find anything with a comparable UX since Twitter shut Dabble DB down.
We're a non-profit project. The core team is based out of a US 501(c)(3).
Features:
* Built on Postgres: Connect to an existing Postgres database or set one up from scratch.
* Utilizes Postgres Features: Mathesar’s UI uses Postgres features. e.g. "Links" in the UI are foreign keys in the database.
* Set up Data Models: Easily create and update Postgres schemas and tables.
* Data Entry: Use our spreadsheet-like interface to view, create, update, and delete table records.
* Data Explorer: Use our Data Explorer to build queries without knowing anything about SQL or joins.
* Schema Migrations: Transfer columns between tables in two clicks in the UI.
* Custom Data Types:: Custom data types for emails and URLs (more coming soon), validated at the database level.
Links:
CODE: https://github.com/centerofci/mathesar
LIVE DEMO: https://demo.mathesar.org/
DOCS: https://docs.mathesar.org/
COMMUNITY: https://wiki.mathesar.org/en/community
WEBSITE: https:/mathesar.org/
SPONSOR US: https://github.com/sponsors/centerofci or https://opencollective.com/mathesar
So happy to realize your stack includes Python, Postgres, Typescript and Svelte because I once flirted with such idea for a personal project. I'm inspired.
If I may ask, have you found using a Typed language such as Typescript has changed your tendency to use Python for other tasks, given that it isn't typed? [0]
[0] - I understand that types can be added on top, but I never found the integration to work that well with mypy when I last tried it many years ago.
I haven't because I'm Python biased. ;) However, for Web I've found myself leaning to TypeScript for type-safety and DX reasons.
Have you ever tried [Pydantic](https://docs.pydantic.dev)!? It may be what you need for type safety / data validation.
JSONLD types are specified with @type, and the range of a @type attribute includes rdfs:Class.
icontract and pycontracts (Design-by-Contract programming) have runtime type and constraint checking; data validation. Preconditions, Command, Postconditions (assertions, assertions of invariance after command C_funcname executed) https://github.com/Parquery/icontract
pydantic_schemaorg: https://github.com/lexiq-legal/pydantic_schemaorg
> Pydantic_schemaorg contains all the models defined by schema.org. The pydantic classes are auto-generated from the schema.org model definitions that can be found on https://schema.org/version/latest/schemaorg-current-https.js... [ https://github.com/schemaorg/schemaorg/tree/main/data/releas... ]
Portable low-field MRI scanners could revolutionize medical imaging
How does this MRI neuroimaging capability differ from openwater's Phase Wave aoproach? https://www.openwater.cc/technology
Is MRI-level neuroimaging possible with just NIRS Near-Infrared Spectroscopy?
First Law of Thermodynamics Breakthrough Upends Equilibrium Theory in Physics
Laws of Thermodynamics (in 2023) https://en.wikipedia.org/wiki/Laws_of_thermodynamics :
> The zeroth law of thermodynamics defines thermal equilibrium and forms a basis for the definition of temperature: If two systems are each in thermal equilibrium with a third system, then they are in thermal equilibrium with each other.
> The first law of thermodynamics states that, when energy passes into or out of a system (as work, heat, or matter), the system's internal energy changes in accordance with the law of conservation of energy.
> The second law of thermodynamics states that in a natural thermodynamic process, the sum of the entropies of the interacting thermodynamic systems never decreases. A common corollary of the statement is that heat does not spontaneously pass from a colder body to a warmer body.
> The third law of thermodynamics states that a system's entropy approaches a constant value as the temperature approaches absolute zero. With the exception of non-crystalline solids (glasses), the entropy of a system at absolute zero is typically close to zero. [2]
First law of thermodynamics https://en.wikipedia.org/wiki/First_law_of_thermodynamics
Quantum thermodynamics https://en.wikipedia.org/wiki/Quantum_thermodynamics
Thermal quantum field theory https://en.wikipedia.org/wiki/Thermal_quantum_field_theory :
> In theoretical physics, thermal quantum field theory (thermal field theory for short) or finite temperature field theory is a set of methods to calculate expectation values of physical observables of a quantum field theory at finite temperature.
From the article:
> [170 years ago] the technology of the time dictated the gases or fluids that people would have studied are in equilibrium at the densities and temperatures that they were using back then.”
"Quantifying Energy Conversion in Higher-Order Phase Space Density Moments in Plasmas" (2023) https://link.aps.org/doi/10.1103/PhysRevLett.130.085201
- "Thermodynamics of Computation Wiki" https://news.ycombinator.com/item?id=18146854
- Phase diagram > Types; where is plasma in this diagram https://en.wikipedia.org/wiki/Phase_diagram#Types
Potential applications?:
- "Thin film" and compact pulsed fusion plasma confinement reactor thermoelectric efficiency ; https://en.m.wikipedia.org/wiki/Helion_Energy
- What do plasmas compute, as a computation medium per Constructor Theory and/or an information medium in deep space?
- Can laser transmutation be achieved more efficiently with the heat of a sustained compact fusion plasma reactor? 4He and 3He are useful outputs with an equilibrium price in recent years, apparently.
There should probably be a facility with omnidirectional conveyors for automated sample testing that runs samples next to the heat?
The Missing Semester of Your CS Education
The course videos: "Missing Semester IAP 2020" https://youtube.com/playlist?list=PLyzOVJj3bHQuloKGG59rS43e2...
Software Carpentry has some similar, OER lessons: https://software-carpentry.org/lessons/
DeepMind has open-sourced the heart of AlphaGo and AlphaZero
Worth noting that while AlphaGo and AlphaZero are incredible achievements, the amount of actual code to implement them isn't very much.
If you have the research paper, someone in the field could reimplement them in a few days.
Then there is the large compute cost for training them to produce the trained weights.
So, opensourcing these bits of work without the weights isn't as major a thing as you might imagine.
> If you have the research paper, someone in the field could reimplement them in a few days.
Hi I did this while I was at Google Brain and it took our team of three more like a year. The "reimplementation" part took 3 months or so and the rest of the time was literally trying to debug and figure out all of the subtleties that were not quite mentioned in the paper. See https://openreview.net/forum?id=H1eerhIpLV
Replication crisis: https://en.wikipedia.org/wiki/Replication_crisis :
> The replication crisis (also called the replicability crisis and the reproducibility crisis) is an ongoing methodological crisis in which the results of many scientific studies are difficult or impossible to reproduce. Because the reproducibility of empirical results is an essential part of the scientific method,[2] such failures undermine the credibility of theories building on them and potentially call into question substantial parts of scientific knowledge.
People should publish automated tests. How does a performance-optimizer know that they haven't changed the output of there are no known-good inputs and outputs documented as executable tests? Pytest-hypothesis seems like a nice compact way to specify tests.
AlphaZero: https://en.wikipedia.org/wiki/AlphaZero
GH topic "AlphaZero" https://github.com/topics/alphazero
I believe ther are one or more JAX implementations of AlphaZero?
Though there's not yet a quantum-inference-based self-play (AlphaZero) algorithm?
TIL about the modified snow plow problem is a variation on TSP, and there are already quantum algos capable of optimally solving TSP.
I think I agree with everything you've said here, but just want to note that while we absolutely should (where relevant) expect published code including automated tests, we should not typically consider reproduction that reuses that code to be "replication" per se. As I understand it, replication isn't merely a test for fraud (which rerunning should typically detect) and mistakes (which rerunning might sometimes detect) but also a test that the paper successfully communicates the ideas such that other human minds can work with them.
Sources of variance; Experimental Design, Hardware, Software, irrelevant environmental conditions/state, Data (Sample(s)), Analysis
Can you run the notebook again with the exact same data sample (input) and get the same charts and summary statistics (output)? Is there a way to test the stability of those outputs over time?
Can you run the same experiment (the same 'experimental design'), ceteris paribus (everything else being equal) and a different sample (input) and get a very similar output? Is it stable, differentiable, independent, nonlinear, reversible; Does it converge?
Now I have to go look up the definitions for Replication, Repeatability, Reproducibility
Replication: https://en.wikipedia.org/wiki/Replication (disambiguation)
Replication_(scientific_method) -> Reproducibility https://en.wikipedia.org/wiki/Reproducibility :
> Measures of reproducibility and repeatability: In chemistry, the terms reproducibility and repeatability are used with a specific quantitative meaning. [7] In inter-laboratory experiments, a concentration or other quantity of a chemical substance is measured repeatedly in different laboratories to assess the variability of the measurements. Then, the standard deviation of the difference between two values obtained within the same laboratory is called repeatability. The standard deviation for the difference between two measurement from different laboratories is called reproducibility. [8] These measures are related to the more general concept of variance components in metrology.
Replication (statistics) https://en.wikipedia.org/wiki/Replication_(statistics) :
> In engineering, science, and statistics, replication is the repetition of an experimental condition so that the variability associated with the phenomenon can be estimated. ASTM, in standard E1847, defines replication as "... the repetition of the set of all the treatment combinations to be compared in an experiment. Each of the repetitions is called a replicate."
> Replication is not the same as repeated measurements of the same item: they are dealt with differently in statistical experimental design and data analysis.
> For proper sampling, a process or batch of products should be in reasonable statistical control; inherent random variation is present but variation due to assignable (special) causes is not. Evaluation or testing of a single item does not allow for item-to-item variation and may not represent the batch or process. Replication is needed to account for this variation among items and treatments.
Accuracy and precision: https://en.m.wikipedia.org/wiki/Accuracy_and_precision :
> In simpler terms, given a statistical sample or set of data points from repeated measurements of the same quantity, the sample or set can be said to be accurate if their average is close to the true value of the quantity being measured, while the set can be said to be precise if their standard deviation is relatively small.
Reproducible builds; to isolate and minimize software variance: https://en.wikipedia.org/wiki/Reproducible_builds
Re: reproducibility, containers, Jupyter books, REES, repo2docker: https://news.ycombinator.com/item?id=32965961 https://westurner.github.io/hnlog/#comment-32965961 (Ctrl-F #linkedreproducibility)
Emacs is on F-Droid
This F-Droid emacs appears to have been build straight from the emacs repo which is cool! I was hoping for a screenshot but I couldn't find one.
I have emacs installed in Termux on Android. Termux provides an extra button bar above the main keyboard for things like Ctrl, arrow keys, Esc, Meta which makes emacs just about usable in Termux. Without those things it would be unusable!
Without the magic button bar you'd need the hackers keyboard which works very well on tablets but it is just a bit squashed for phones unless you use them in landscape.
https://play.google.com/store/apps/details?id=org.pocketwork...
For good measure, let's share the link to Hacker's Keyboard on F-droid while we are at it :-)
https://f-droid.org/packages/org.pocketworkstation.pckeyboar...
Here's how to install conda, mamba, and pip with MambaForge in Termux w/ proot from FDroid because there are no official APKs on the play store as is now necessary to bless binaries with context labels: https://github.com/westurner/dotfiles/blob/develop/scripts/s...
Physicists Use Quantum Mechanics to Pull Energy Out of Nothing
Energy teleportation in 2023:
> Now in the past year, researchers have teleported energy across microscopic distances in two separate quantum devices, vindicating Hotta’s theory. The research leaves little room for doubt that energy teleportation is a genuine quantum phenomenon."
Quantum foam: https://en.wikipedia.org/wiki/Quantum_foam
Vikings went to Mediterranean for ‘summer jobs’ as mercenaries, left graffiti
Would you put graffiti on your own vehiclr?
Careers for graphic artists with BA degrees
Prehistoric art: https://en.wikipedia.org/wiki/Prehistoric_art
1. "Encino Man" (1992) https://en.wikipedia.org/wiki/Encino_Man https://www.google.com/search?kgmid=/m/05fc8m&hl=en-US&q=Enc...
Social media is a cause, not a correlate, of mental illness in teen girls
Perhaps Facebook should again require an email address at an approving college or university for sign up.
But what about supporting the NCMEC National Center for Missing & Exploited Children, without biometrics due to new laws?
Can you teach your daughters to be considerate of others on the internet?
They used to tell us not to put any personal information online as kids; no real names, etc.
Perhaps social media is a mirror from which we can determine parenting approach?
Are violent video games less bad than social media for which unhealthily-competitive strata?
"Please do not bully people on the internet because:"
"Please be considerate and helpful on the internet because:"
StopBullying.gov: https://www.stopbullying.gov/
"How to Make a Family Media Use Plan" (HealthyChildten.org (AAP)) https://www.healthychildren.org/English/family-life/Media/Pa...
Internet safety: https://en.wikipedia.org/wiki/Internet_safety
[after school] emotional restraint collapse: https://www.google.com/search?q=emotional+restraint+collapse :
; It's exhausting for many of us to withhold nonverbal emotions all day; and so after school a minute to just chill without typical questioning may or may not help prevent bullying on the internet
TIL Iceland provides vouchers for whatever after school activities: taxes pay for sports and other programs.
The fundamental thermodynamic costs of communication
I like how "Due to the generality of our model, our findings could help explain empirical observations of how thermodynamic costs of information transmission make inverse multiplexing energetically favorable in many real-world communication systems." makes it sound like biology is not the real world
Took me a while to understand this comment. I'm not a biologist.
A nerve, is a whole big bundle of axons[1]. Looks like each bundle carries hundreds or thousands of connections down to a muscle.
1 http://www.medicine.mcgill.ca/physio/vlab/other_exps/CAP/ner...
I am a neurobiologist and still don't understand the comment. The authors acknowledge biology in the first sentence.
wrt. claims in the full article however I have an issue with this observation...
> Another feature common to many biological and artificial communication systems is that they split a stream of information over multiple channels, i.e., they inverse multiplex. Often this occurs even when a single one of the channels could handle the entire communication load. For example, multiple synapses tend to connect adjacent neurons and multiple neuronal pathways tend to connect brain regions.
I dont see how a "single channel" i.e. a single synapse would suffice as the entire information output unit of a neuron. Making multiple synaptic connections is fundamental to the way neural networks perform computations.
Furthermore, representational drift observes that a biological single neuron's output given activation is not stable over time; which implies that there is greater emergent complexity than is modeled with ANNs (which have stable outputs given training, NN topology parameters, and activation functions that effectively weight training samples (which are usually also noise))
/? Representational drift brain https://www.google.com/search?q=representational+drift+brain ...
"Causes and consequences of representational drift" (2019) https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7385530/
> The nervous system learns new associations while maintaining memories over long periods, exhibiting a balance between flexibility and stability. Recent experiments reveal that neuronal representations of learned sensorimotor tasks continually change over days and weeks, even after animals have achieved expert behavioral performance. How is learned information stored to allow consistent behavior despite ongoing changes in neuronal activity? What functions could ongoing reconfiguration serve? We highlight recent experimental evidence for such representational drift in sensorimotor systems, and discuss how this fits into a framework of distributed population codes. We identify recent theoretical work that suggests computational roles for drift and argue that the recurrent and distributed nature of sensorimotor representations permits drift while limiting disruptive effects. We propose that representational drift may create error signals between interconnected brain regions that can be used to keep neural codes consistent in the presence of continual change. These concepts suggest experimental and theoretical approaches to studying both learning and maintenance of distributed and adaptive population codes.
"The geometry of representational drift in natural and artificial neural networks" (2022) https://journals.plos.org/ploscompbiol/article?id=10.1371/jo... :
> [...] We examine stimulus representations from fluorescence recordings across hundreds of neurons in the visual cortex using in vivo two-photon calcium imaging and we corroborate previous studies finding that such representations change as experimental trials are repeated across days. This phenomenon has been termed “representational drift”. In this study we geometrically characterize the properties of representational drift in the primary visual cortex [...]
> The features we observe in the neural data are similar to properties of artificial neural networks where representations are updated by continual learning in the presence of dropout, i.e. a random masking of nodes/weights, but not other types of noise. Therefore, we conclude that a potential reason for the representational drift in biological networks is driven by an underlying dropout-like noise while continuously learning and that such a mechanism may be computational advantageous for the brain in the same way it is for artificial neural networks, e.g. preventing overfitting.
"Neurons are fickle: Electric fields are more reliable for information" (2022) https://www.sciencedaily.com/releases/2022/03/220311115326.h... :
> [...] And when the scientists trained software called a "decoder" to guess which direction the animals were holding in mind, the decoder was relatively better able to do it based on the electric fields than based on the neural activity.
> This is not to say that the variations among individual neurons is meaningless noise, Miller said. The thoughts and sensations of people and animals experience, even as they repeat the same tasks, can change minute by minute, leading to different neurons behaving differently than they just did. The important thing for the sake of accomplishing the memory task is that the overall field remains consistent in its representation.
> "This stuff that we call representational drift or noise may be real computations the brain is doing, but the point is that at that next level up of electric fields, you can get rid of that drift and just have the signal," Miller said.
> The researchers hypothesize that the field even appears to be a means the brain can employ to sculpt information flow to ensure the desired result. By imposing that a particular field emerge, it directs the activity of the participating neurons.
> Indeed, that's one of the next questions the scientists are investigating: Could electric fields be a means of controlling neurons?
/? representational drift site:github.com https://www.google.com/search?q=representational+drift+site%...
Computational neuroscience: https://en.wikipedia.org/wiki/Computational_neuroscience :
> Models in theoretical neuroscience are aimed at capturing the essential features of the biological system at multiple spatial-temporal scales, from membrane currents, and chemical coupling via network oscillations, columnar and topographic architecture, nuclei, all the way up to psychological faculties like memory, learning and behavior. These computational models frame hypotheses that can be directly tested by biological or psychological experiments.
Are you making some sort of point, or just dumping content?
Perhaps I was too polite. The collapsed entropy (absent real world noise per observation) of the binary relations in the brain is a useful metric.
Interesting take on being impolite - making a basic statement using overtly pretensions and opaque language?
> [...] Here we present the first study that rigorously combines such a framework, stochastic thermodynamics, with Shannon information theory. We develop a minimal model that captures the fundamental features common to a wide variety of communication systems. We find that the thermodynamic cost in this model is a convex function of the channel capacity, the canonical measure of the communication capability of a channel. We also find that this function is not always monotonic, in contrast to previous results not derived from first principles physics. These results clarify when and how to split a single communication stream across multiple channels. In particular, we present Pareto fronts that reveal the trade-off between thermodynamic costs and channel capacity when inverse multiplexing. Due to the generality of our model, our findings could help explain empirical observations of how thermodynamic costs of information transmission make inverse multiplexing energetically favorable in many real-world communication systems.
https://arxiv.org/abs/2302.04320
What is the Shannon entropy interpretation of e.g. (quantum wave function) amplitude encoding?
"Quantum discord" https://en.wikipedia.org/wiki/Quantum_discord
> In quantum information theory, quantum discord is a measure of nonclassical correlations between two subsystems of a quantum system. It includes correlations that are due to quantum physical effects but do not necessarily involve quantum entanglement.
Isn't there more entropy if we consider all possible nonlocal relations between bits; or, is which entropy metric independent of redundant coding schemes between points in spacetime?
New neural network architecture inspired by neural system of a worm
It makes a good headline, but reading over the paper (https://www.nature.com/articles/s42256-022-00556-7.pdf) it doesn’t seem biologically-inspired. It seems like they found a way to solve nonlinear equations in constant time via an approximation, then turned that into a neural net.
More generally, I’m skeptical that biological systems will ever serve as a basis for ML nets in practice. But saying that out loud feels like daring history to make a fool of me.
My view is that biology just happened to evolve how it did, so there’s no point in copying it; it worked because it worked. If we have to train networks from scratch, then we have to find our own solutions, which will necessarily be different than nature’s. I find analogies useful; dividing a model into short term memory vs long term memory, for example. But it’s best not to take it too seriously, like we’re somehow cloning a brain.
Not to mention that ML nets still don’t control their own loss functions, so we’re a poor shadow of nature. ML circa 2023 is still in the intelligent design phase, since we have to very intelligently design our networks. I await the day that ML networks can say “Ok, add more parameters here” or “Use this activation instead” (or learn an activation altogether — why isn’t that a thing?).
"I’m skeptical that biological systems will ever serve as a basis for ML nets in practice"
First of all, ML engineers need to stop being so brainphiliacs, caring only about the 'neural networks' of the brain or brain-like systems. Lacrymaria olor has more intelligence, in terms of adapting to exploring/exploiting a given environment, than all our artificial neural networks combined and it has no neurons because it is merely a single-cell organism [1]. Once you stop caring about the brain and neurons and you find out that almost every cell in the body has gap junctions and voltage-gated ion channels which for all intents and purposes implement boolean logic and act as transistors for cell-to-cell communication, biology appears less as something which has been overcome and more something towards which we must strive with our primitive technologies: for instance, we can only dream of designing rotary engines as small, powerful, and resilient as the ATP synthase protein [2].
[1] Michael Levin: Intelligence Beyond the Brain, https://youtu.be/RwEKg5cjkKQ?t=202
[2] Masasuke Yoshida, ATP Synthase. A Marvellous Rotary Engine of the Cell, https://pubmed.ncbi.nlm.nih.gov/11533724
> Once you stop caring about the brain and neurons and you find out that almost every cell in the body has gap junctions and voltage-gated ion channels which for all intents and purposes implement boolean logic and act as transistors for cell-to-cell communication, biology appears less as something which has been overcome and more something towards which we must strive with our primitive technologies: for instance, we can only dream of designing rotary engines as small, powerful, and resilient as the ATP synthase protein [2].
But what of wave function(s); and quantum chemistry at the cellular level? https://github.com/tequilahub/tequila#quantumchemistry
Is emergent cognition more complex than boolean entropy, and are quantum primitives necessary to emulate apparently consistently emergent human cognition for whatever it's worth?
[Church-Turing-Deutsch, Deutsch's Constructor theory]
Is ATP the product of evolutionary algorithms like mutation and selection? Heat/Entropy/Pressure, Titration/Vibration/Oscillation, Time
From the article:
> The next step, Lechner said, “is to figure out how many, or how few, neurons we actually need to perform a given task.”
Notes regarding Representational drift* and remarkable resilience to noise in BNNs) from "The Fundamental Thermodynamic Cost of Communication: https://news.ycombinator.com/item?id=34770235
It's never just one neuron.
And furthermore, FWIU, human brains are not directed graphs of literally only binary relations.
In a human brain, there are cyclic activation paths (given cardiac electro-oscillations) and an imposed (partially extracerebral) field which nonlinearly noises the almost-discrete activation pathways and probably serves a feed-forward function; and in those paths through the graph, how many of the neuronal synapses are simple binary relations (between just nodes A and B)?
> The group also wants to devise an optimal way of connecting neurons. Currently, every neuron links to every other neuron, but that’s not how it works in C. elegans, where synaptic connections are more selective. Through further studies of the roundworm’s wiring system, they hope to determine which neurons in their system should be coupled together.
Is there an information metric which expresses maximal nonlocal connectivity between bits in a bitstring; that takes all possible (nonlocal, discontiguous) paths into account?
`n_nodes*2` only describes all of the binary, pairwise possible relations between the bits or qubits in a bitstring?
"But what is a convolution" https://www.3blue1brown.com/lessons/convolutions
Quantum discord: https://en.wikipedia.org/wiki/Quantum_discord
Scientists find first evidence that black holes are the source of dark energy
Introduction to Datalog
In the last few months the mention of Datalog has increased, I wondered how it differed from graph databases and found a clear answer in SO [1]. I am not an incumbent but found graph databases and clause approaches interesting.
[1] https://stackoverflow.com/questions/29192927/a-graph-db-vs-a... (2015)
I did an interview [1] with Kevin Feeney, one of the founders (no longer active) of TerminusDb, which goes into some depth about the difference between RDF stores and (property) graph databases, where the former is more closely aligned with datalog and logic programming. There are links to a really excellent series of blog posts by Kevin on this topic in the show notes.
[1] https://thesearch.space/episodes/5-kevin-feeney-on-terminusd...
With RDF* and SPARQL* ("RDF-star" and "SPARQL-star") how are triple (or quad) stores still distinct from property graphs?
RDFS and SHACL (and OWL) are optional in a triple store, which expects the subject and predicate to be string URIs, and there is an object datatype and optional language:
(?s ?p ?o <datatype> [lang])
(?subject:URI, ?predicate:URI, ?object:datatype, object_datatype, [object_language])
RDFS introduces rdfs:domain and rdfs:range type restrictions for Properties, and rdfs:Class and rdfs:subClassOf.`a` means `rdf:type`; which does not require RDFS:
("#xyz", a, "https://schema.org/Thing")
("#xyz", rdf:type, "https://schema.org/Thing")
Quad stores have a graph_id string URI "?g" for Named Graphs: (?g ?s ?p ?o)
("https://example.org/ns/graphs/0", "#xyz", a, "https://schema.org/Thing")
("https://example.org/ns/graphs/1", "#xyz", a, "https://schema.org/ScholarlyArticle")
There's a W3C CG (Community Group) revising very many of the W3C Linked Data specs to support RDF-star: https://www.w3.org/groups/wg/rdf-starLooks like they ended up needing to update basically most of the current specs: https://www.w3.org/groups/wg/rdf-star/tools
"RDF-star and SPARQL-star" (Draft Community Group Report; 08 December 2022) https://w3c.github.io/rdf-star/cg-spec/editors_draft.html
GH topics: rdf-star, rdfstar: https://github.com/topics/rdf-star, https://github.com/topics/rdfstar
pyDatalog does datalog with SQLAlchemy and e.g. just the SQLite database: https://github.com/pcarbonn/pyDatalog ; and it is apparently superseded by IDP-Z3: https://gitlab.com/krr/IDP-Z3/
From https://twitter.com/westurner/status/1000516851984723968 :
> A feature comparison of SQL w/ EAV, SPARQL/SPARUL, [SPARQL12 SPARQL-star, [T-SPARQL, SPARQLMT,]], Cypher, Gremlin, GraphQL, and Datalog would be a useful resource for evaluating graph query languages.
> I'd probably use unstructured text search to identify the relevant resources first.
Show HN: Polymath: Convert any music-library into a sample-library with ML
Polymath is a open-source tool that converts any music-library into a sample-library with machine learning. It separates songs into stems, quantizes to same BPM, detects key and much more. A game-changing workflow for music producers & DJ
Other cool #aiart things:
- "MusicLM: Generating music from text" https://news.ycombinator.com/item?id=34543748
- /? awesome Generative AI site:GitHub.com https://www.google.com/search?q=awesome+generative+ai+site%3...
- #GenerativeArt #GenerativeMusic
- BespokeSynth DAW: https://github.com/BespokeSynth/BespokeSynth :
> [...] live-patchable environment, so you can build while the music is playing; VST, VST3, LV2 hosting; Python livecoding; MIDI, [...]
> Using Polymath's search capability to discover related tracks, it is a breeze to create a polished, hour-long mash-up DJ set
https://github.com/samim23/polymath#how-does-it-work :
> How does it work?
> - Music Source Separation is performed with the [facebook/demucs] neural network
> - Music Structure Segmentation/Labeling is performed with the [wayne391/sf_segmenter] neural network
> - Music Pitch Tracking and Key Detection are performed with [marl/Crepe] neural network
> - Music Quantization and Alignment are performed with [bmcfee/pyrubberband]
> - Music Info retrieval and processing is performed with [librosa/librosa]
Coffee won’t give you extra energy, just borrow a bit that you’ll pay for later
From the article:
> This is because the caffeine won’t bind forever, and the adenosine that it blocks doesn’t go away. So eventually the caffeine breaks down, lets go of the receptors and all that adenosine that has been waiting and building up latches on and the drowsy feeling comes back – sometimes all at once.
> So, the debt you owe the caffeine always eventually needs to be repaid, and the only real way to repay it is to sleep.
Does drinking water offset the exertion-resultant dehydration that caffeine and other stimulants tend to result in?
[dead]
Let Teenagers Sleep
As a teenager I was diagnosed with "ADD". Very intelligent, but unable to focus or complete assignments.
My life habits were carb heavy unhealthy food (from the cafeteria), soda, lack of sleep due to long school commute, not much exercise
As an adult, I eat no carbs, all meat and vegetables, I work from home and sleep in as far as I need every night. My thinking is laser sharp.
They tried to medicate me with all sorts of anti depression drugs and amphetamines. Turns out I was just very unhealthy, from a basic lifestyle perspective. And the school was pushing that lifestyle. My guidance counseler suggested I dont attend college, just a community college (despite the fact that I got admitted to a decent state school), or maybe go into the "trades".
These large scale school systems treat students like cattle, with zero regard for the long term effects. Many mental health concerns would disappear if people were actually healthy in the most basic sense.
/? ADHD and sleep; REM / non-REM: https://www.google.com/search?q=adhd+and+sleep
Sleep hygiene: https://en.wikipedia.org/wiki/Sleep_hygiene https://www.google.com/search?q=sleep+hygiene
- Enough exercise and water
- Otherwise, limit calories and/or protein in the preceding hours
Sleep induction: https://en.wikipedia.org/wiki/Sleep_induction
Pranayama; breathing yoga: https://en.wikipedia.org/wiki/Pranayama
4-7-8 breathing: https://www.google.com/search?q=4-7-8+breathing
Attributed both to Army and Navy: https://www.fastcompany.com/90253444/what-happened-when-i-tr... :
> The Independent says the technique was first described in a book from 1981 called "Relax and Win: Championship Performance" by Lloyd Bud Winter.
/? "Relax and Win: Championship Performance" https://www.google.com/search?q=%22Relax+and+Win%3A+Champion... https://archive.org/details/Relax-and-Win_Championship-Perfo...
Why is there so much useless and unreliable software?
Linear logic has been known since 1987. The first release of Coq (dependent types for functional programming and writing proofs) was in 1989. The HoTTBook came out in 2013. Ada/SPARK 2014 came out the same year as Java 8 did. We also witnessed the Software Foundations series, the CompCert C compiler, the Sel4 microkernel, and the SPARKNaCl cryptographic library.
Instead of learning about those achievements and aiming to program for the same reliability, clarity, and sophistication, we see an abundance of software that cannot clearly describe their own behavior nor misbehavior.
Instead of incorporating the full functionality of XML/HTML/CSS/SVG/JS/WebGL into the development experience and providing ways to control them at the fundamental level, we reinvent crude approximations like the various web frameworks.
YAML and JSON often trumps XML/XSD until things get out of control, and even then, people still don't learn the lesson. Protobuf, flatbuffer, capnproto, and the like keep reinventing ASN.1.
Naive microservices partially reimplements Erlang's BEAM VM while ignoring all the hard parts that BEAM VM got right. Many people riding the microservice bandwagon have never even heard of Paxos, not to mention TLA+.
Many programmers keep learning new shining frameworks but are reluctant to learn about the crucial fundamentals, e.g., Introduction to Parallel Algorithms and Architectures, nor how to think clearly and unambiguously in the spirit of Coq/Agda/Lean.
No wonder ChatGPT exposes how shallow most of programming is and how lacking most programmers are in actual understanding. Linear logic and dependent types are there to help us design and think with clarity at a high level, but people would rather fumble around with OOP class hierarchies (participate in the pointless is-a/has-a arguments) and "architecture" design that only complicate things.
What is this madness? This doesn't sound like engineering.
The adoption curve of advanced technologies that solve it all. Is it just the cognitive burden of the tooling or the concepts, are the training methods different, a dearth of already trained talent and sufficient post-graduate or post-doctoral instructors, unclear signals between demand and a supply funnel?
The limited availability of Formal Methods and Formal Verification training.
The growing demand for safety critical software for things that are heavy and that move fast and that fly over our heads and our homes.
In order to train, you need to present URLs: CreativeWork(s) in an outline (a tree graph), identify competencies (typically according to existing curricula) and test for comprehension, and what else is necessary to boost retention for and of identified competencies?
There are many online courses, but so few on Computer Science Education. You can teach with autograded Jupyter notebooks and your own edX instance all hosted in OCI containers in your own Kubernetes cloud, for example. Containers and Ansible Playbooks help minimize cost and variance in that part of the learning stack.
We should all learn Coq and TLA+ and Lean, especially. What resources and traversal do you recommend for these possibly indeed core competencies? For which domains are no-code tools safe?
If we were to instead have our LLM (,ChatGPT,Codex,) filter expression trees in order to autocomplete from only Formally Verified code with associated Automated Tests, and e.g. Lean Mathlib, how would our output differ from that of an LLM training on code that may or may not have any tests?
Could that also implement POSIX and which other interfaces, please?
> The adoption curve of advanced technologies that solve it all. Is it just the cognitive burden of the tooling or the concepts, are the training methods different, a dearth of already trained talent and sufficient post-graduate or post-doctoral instructors, unclear signals between demand and a supply funnel?
> The limited availability of Formal Methods and Formal Verification training.*
From "Are software engineering “best practices” just developer preferences?" https://news.ycombinator.com/item?id=28709239 :
>>> From "Why Don't People Use Formal Methods?" https://news.ycombinator.com/item?id=18965964 :
>>>> Which universities teach formal methods?
>>>> - q=formal+verification https://www.class-central.com/search?q=formal+verification
>>>> - q=formal+methods https://www.class-central.com/search?q=formal+methods
>>>> Is formal verification a required course or curriculum competency for any Computer Science or Software Engineering / Computer Engineering degree programs? https://news.ycombinator.com/item?id=28513922
Formal methods should be required course or curriculum competency for various Computer Science and Software Engineering credentials.
> The growing demand for safety critical software for things that are heavy and that move fast and that fly over our heads and our homes.
"The Case of the Killer Robot"
> In order to train, you need to present URLs: CreativeWork(s) in an outline (a tree graph), identify competencies (typically according to existing curricula) and test for comprehension, and what else is necessary to boost retention for and of identified competencies?
A Multi-track "ThingSequence" which covers; And there are URIs for the competencies and exercises that cover.
> There are many online courses, but so few on Computer Science Education. You can teach with autograded Jupyter notebooks and your own edX instance all hosted in OCI containers in your own Kubernetes cloud, for example. Containers and Ansible Playbooks help minimize cost and variance in that part of the learning stack.
Automation with tooling is necessary to efficiently compete. In order to support credentialing workflows that qualify point-in-time performance, it is advisable to adopt tools that support automation of grading, for example.
> We should all learn Coq and TLA+ and Lean, especially. What resources and traversal do you recommend for these possibly indeed core competencies?
Links above. When should Formal Methods be introduced in a post-secondary or undergraduate program? Is there any reason that a curriculum can't go from math proofs, to logical proofs, to logical argumentation?
> For which domains are no-code tools safe?
Unfortunately, no-code tools in even low-risk applications can be critical vulnerabilities if persons do not understand their limitations. For example, "we used to have already-printed [offline] forms on hand [before the new computerized workflows [enabled by no-code, no review development workflows]]".
Should there be 2 or 3 redundant processes with an additional component that discards low-level outputs if there is no consensus? Is that plus No-code tools safe?
> If we were to instead have our LLM (,ChatGPT,Codex,) filter expression trees in order to autocomplete from only Formally Verified code with associated Automated Tests, and e.g. Lean Mathlib, how would our output differ from that of an LLM training on code that may or may not have any tests?
You should not be autocompleting from a training corpus of code without tests.
> Could that also implement POSIX and which other interfaces, please?
Eventually, AI will produce an OS kernel that is interface compatible with what we need to run existing code.
How can experts assess whether there has been sufficient review of an [AI-generated] alternative interface implementation?
Not everybody should learn TLA+. It's a pretty niche tool and it's not helpful for a lot of kinds of software!
What are some of the limits of TLA+ for dynamic analysis? Which other tools model variable-latency distributed systems?
I don't think that most software is a variable-latency distributed system. I wouldn't use it for a simulation, for example, or anything which depends at a high level on non-integer numbers (TLA+ can't model-check floats or reals).
The Rust Implementation of GNU Coreutils Is Becoming Remarkably Robust
I wonder why not reimplement coreutils as library functions, to be used within an ad-hoc REPL. It would be much more flexible and extensible.
AFAIK, originally the reason why they were made as programs was that the available programming languages were too cumbersome for such use. Now we have plenty of experience making lightweight scripting languages that are pleasant to use in a live environment (vs premade script), so why give up the flexibility of ad-hoc scripting?
Isn't what you're describing basically a unix "shell" like bash, csh, zsh, etc?
I agree that something like bash leaves a lot to be desired in REPL functionality. Its ubiquity is convenient, however.
DIY 1,500W solar power electric bike (2022)
It's RadWagon 4 with white paint job, upgraded brakes/power controller, and solar panel at the top of the stock roof.
Not quite what I expected when I read the title.
Btw, the panel is 50W, and I assume the battery is stock, 672 Wh. That's 14+ hours of best-case sunlight to recharge fully. I wonder how practical is it.
Yeah, makes more sense to leave the solar panel in your backyard (saving some weight) and have spare battery always sitting on the solar charger.
Then you can bring your bike inside, in the shade, etc. — not having to always find sunlight to park in.
The only way bringing the solar panel with the bike makes sense is a multi-day ride. But as is being pointed out, you're going to be spending a lot more time sunbathing rather than biking under power.
<wagon in tow> Yeah, Good call boss.
The cost figures in this article about [rooftop,] wind do not take into account latest gen [Dyneema] ultralight rooftop solar:
"Rooftop wind energy innovation claims 50% more energy than solar at same cost" (2022) https://pv-magazine-usa.com/2022/10/14/rooftop-wind-energy-i...
> The scalable, “#motionless” #WindEnergy unit can produce 50% more energy than rooftop solar at the same cost, said the company.
> The technology leverages aerodynamics similar to #airfoils in a race car to capture and amplify each building’s airflow. The unit requires about 10% of the space required by solar panels and generates round-the-clock energy. Aeromine said unlike conventional wind turbines that are noisy, visually intrusive, and dangerous to migratory birds, the patented system is motionless and virtually silent.
> An #Aeromine system typically consists of 20 to 40 units installed on the edge of a building facing the predominant wind direction. The company said the unit can minimize energy storage capacity needed to meet a building’s energy needs, producing energy in all weather conditions. With a small footprint on the roof, the unit can be combined with rooftop solar, providing a new tool in the toolkit for decarbonization and energy independence.
"18 Times More Power: MIT Researchers Have Developed Ultrathin Lightweight Solar Cells" (2022) https://scitechdaily.com/18-times-more-power-mit-researchers... :
> When they tested the device, the MIT researchers found it could generate 730 watts of power per kilogram when freestanding and about 370 watts-per-kilogram if deployed on the high-strength Dyneema fabric, which is about 18 times more power-per-kilogram than conventional solar cells.
> “A typical rooftop solar installation in Massachusetts is about 8,000 watts. To generate that same amount of power, our fabric photovoltaics would only add about 20 kilograms (44 pounds) to the roof of a house,” he says.
> They also tested the durability of their devices and found that, even after rolling and unrolling a fabric solar panel more than 500 times, the cells still retained more than 90 percent of their initial power generation capabilities.
E.g. Hyperlite Mountain Gear sells Dyneema ultralight backpacking packs and coats. There are Dyneema Patch Kits that work for various types of gear.
Wise to look at Ultralight backpacking gear before buying regular camping gear. Solarcore Aerogel is warm and light and also in encased in PVA foam rubber which is like a new wet suit. https://twitter.com/westurner/status/1600820322567041024 Kayaking bags are waterproof, but are there yet Dyneema ones?
You mightn't have understood?
730-370 watts/kilogram is the number to beat (for DIY electric bicycle applications)
And rooftop wind is competitive (for charging offline batteries)
Presumably, bicycling is like ultralight hiking: wHr/kg is the or a limit https://en.wikipedia.org/wiki/Kilowatt-hour
A pedaling electric bicycler could tow a solar wagon, eh
Actors Say They’re Being Asked to Sign Away Their Voice to AI
This is starting to sound eerily similar to that BoJack Horseman plotline, where he won an Oscar for a movie he never acted in, because he signed away his likeness and the director just used a 3d model of him because he was always too drunk to be on set.
But in all honesty though, this could be great news - imagine not needing to do recall shoots because the dialogue was bad, and you can fill in a few sentences with this technology instead. Of course, it will be abused and used to not compensate actors for their work properly.
But if it develops rapidly I could imagine games and animation using "stock voices" in their productions, just like designers often use stock imagery. I'm sure someone is happy to sign away their voice for a few grand, and provide it to a stock voice library.
S1M0̸NE (2002) https://en.wikipedia.org/wiki/Simone_(2002_film) Frankenstein.
Her (2013) https://en.wikipedia.org/wiki/Her_(film) She's too smart for you anyway.
Upload (2020, 2022) https://en.wikipedia.org/wiki/Upload_(TV_series) She buys in game credits for his existence.
Faust / The Little Mermaid: https://en.wikipedia.org/wiki/Faust#Cinematic_adaptations
"Final cut, Suit!" - Billy Walsh, Entourage https://en.wikipedia.org/wiki/List_of_recurring_Entourage_ch...
Final cut privilege: https://en.wikipedia.org/wiki/Final_cut_privilege :
> typically reluctant
Why do we create modern desktop GUI apps using HTML/CSS/JavaScript? (2022)
Because Web Standards are Portable and Accessible.
You can recreate the accessible GUI widget tree universe in ASM or WASM, but it probably won't be as accessible as standard HTML form elements unless you spend more time than you have for that component on it.
For example, video game devs tend to do this: their very own text widget (and hopefully a ui scaling factor) with a backgroundColor attribute - instead of CSS's background-color - and then it doesn't support tabindex or screen readers or high contrast mode or font scaling.
It's a widget tree with events either way, but Web Standards are Portable and Accessible (with a comparative performance cost that's probably with it)
> Because Web Standards are Portable and Accessible.
A good http-friendly GUI standard would be also (as I describe nearby). But do note that businesses still use mostly desktops for everyday CRUD and don't want to pay a large "mobile tax" if given a choice. YAGNI has been ignored, as the standards over-focused on social media and e-commerce at the expense of typical CRUD. CRUD ain't sexy but necessary and common.
It's not easy to do both desktop and mobile UI's well and inexpensively in one shot. I believe there are either inherent trade-offs between them, or that the grand unification UI framework has yet to be invented. (At least one that doesn't require UI rocket science.)
Flutter > See also is a good list: https://en.wikipedia.org/wiki/Flutter_(software)
/? flutter python ... TIL about flet: https://github.com/flet-dev/flet https://flet.dev/docs/guides/python/getting-started
Mobile-first development says develop the mobile app first and teh desktop version can get special features later; one responsive layout for Phone, Tablet, and desktop
Phosh (GTK) and KDE Plasma Mobile are alternatives to iOS and Android (which do have a terminal, bash, git, and a way to install CPython (w/ MambaForge ARM64 packages) and IPython)
Ask HN: Advice from people who strength train from home
I have a pullup bar and a Lebert Equalizer kind of thing. I live in a small room at my university. I am planning on training bodyweight or calisthenics as it is called popularly.
HNers who train from home using minimal weights or equipments, can you suggest a path for me.
I am looking for some hints on:
1. What is the bare minimum balanced routine I can start with?
2. How long should I stick to it before I see or feel actual results from it?
3. Diet? Eggs are easily available for me.
4. A bit about your journey. How you started and how have you progressed on parameters of strength, routine, size, energy, etc.
P.S: I came across this youtube channel: https://www.youtube.com/@Kboges where he suggests that to gain strength and general fitness you can train daily with 3 movements but not to failure. Is it possible?
My goals are to have enough muscle and strength so that I don't get tired doing chores lifting something for my household. I want this to go far into my old age so that I don't fall and spend my final years in a nursing home bed.
(Secondhand) Total Gym XLS & 4x 20lb 5gal bucket concrete weights on 1.25" fittings, Basketball filled with sand, occasional Yoga, lately very occasional Inline Skating up and down a hill with wrist guards and a MIPS helmet with a visor
(Soccer, then former middle of the pack Distance Running; then also weight training resistance training in a Weider cage with an 45lb bar and adjustable spotter bars, high pulley, low pulley, leg extension; then Bowflex; and now TotalGym and I prefer it. Lol, you watch the infomercial and you see the testimonials and you think "nobody's that happy with their without ever" and still I really do enjoy this equipment. (I have never been paid to endorse any fitness product or book.))
Calisthenics says that "time under tension" is more relevant than number of reps.
The TotalGym is a decent to good partner stretcher. "Don't slam the stack, and don't waste an opportunity to let it stretch you out"
Dance music has a higher tempo. Various apps will generate workout playlists with AFAIU a BPM ladder
The "high protein foods" part of Keto.
My understanding of Keto: if you eat too much sugar (including starchy carbs) without protein, the body learns to preferentially burn sugar and wastes the protein; so eat protein all day. We do need carbs to efficiently process protein. (And we do need fat for our brains: there is a reference DV daily value for fat for a reason. Ranch, Whole Milk, and Peanut Butter have fat.)
Omega-3s and Omega-6s would be listed under "Polyunsaturated Fats" if it were allowed to list them on the standard Nutrition Facts label instead of elsewhere on the packaging.
Omega 6:3 balance apparently affects endocannabinoid levels. Endocannabinoids help regulate diet and inflammation.
Antioxidants: Vitamins A, C, E,
Electrolytes: Water + Salt + Potassium (H20 + NaCl + K)
There are many lists of foods to eat for inflammation and inflammatory conditions.
Foods rich in anthocyanins tend to be high in nutrients; for example, blueberries, chard, and other dark leafy vegetables and fruits. Blueberry smoothie; premixed, frozen, fresh and washed with sprayed water+vinegar in a clip-on colander.
Pressure cooking is a relatively healthy way to prepare protein.
You can make a dozen hard-boiled or soft-boiled eggs with one instant pot pressure cooker in ~20 minutes and with less water than conventional boiling.
50g/day of Protein is 100% of the DV for a 2000 calorie diet, when you're not trying to gain muscle mass. Bodybuilders consume at least 100g or 150g of protein a day.
Tired of e.g. tuna, eggs, beef; I learned of "The No Meat Athlete Cookbook: Whole Food, Plant-Based Recipes to Fuel Your Workouts—and the Rest of Your Life", which has a bunch of ideas for vegan and vegetarian protein and nutrionally-balanced meals.
FWIU, vegan and vegetarian diets tend to have some common issues like e.g. magnesium deficiency.
I can't be mad at the lion for being omnivorous; but frustrated at the lion for being greedy and selfish in regards to ecology and smell. /? S tool reading infographic
Protein bars (20g: Huel, Aldi), Protein bread (10g/slice), Muscle Milk Pro (50g), Huel, Filtered water: Wide-mouth water bottles, Fish Oil (Omega-3s DHA & EPA), Olive Oil (<~380°F), spray Avocado Oil (Aldi), Peanut Butter, Whole Milk, Multigrain Cheerios over Total (in the 100% DV cereals category),
Supergrains: flax, chia, shelled hemp seed. Super grains mixed into peanut butter = $25 health food store peanut butter.
Healthy Eating Plate: Water, Fruits, Grains, Veggies, Protein. USDA Myplate: people need* Milk/Dairy (which is indeed basically impossible to create a synthetic analogue of, in terms of formula or)
Some greens have little more nutritional content than water and fiber. AFAIU, Chard is as nutritious as Spinach, which has iron (which is what Popeye eats to hopefully eventually woo Olive Oil)
Ice water diminishes appetite. Bread has filling carbs that you can eat with protein* (to stay closer to "ketosis", for example)
HIIT says don't rest for more than 15 seconds / 45 seconds between exercises / sets
My TotalGym workouts now are much more aerobic than when I started getting back to healthy and counted reps. I've put off adding more weight to the carriage bar (that certain TG models lierally support) for like a year and I've rounded-out in areas I mightn't have as a result.
I don't miss free weights, stacks / universal machines. After trying rings with nobody else around in the backyard, the TotalGym wins.
Certain (my parents') Bowflex units can't be upgraded with more or heavier tension rods; though the weighted non-inclined rowing is cool too.
Watched a few "Bodybuilding on a budget" [how to buy protein at at discount grocery store] yt videos. I usually eat cold, but Instant Pot for the win; TIL grill char is carcinogenic. With a second Instant Pot, you can do veggies separately from protein, which takes much longer to cook.
Show HN: DocsGPT, open-source documentation assistant, fully aware of libraries
Hi, This is a very early preview of a new project, I think it could be very useful. Would love to hear some feedback/comments
From https://news.ycombinator.com/item?id=34659668 :
>> How do the responses compare to auto-summarization in terms of Big E notation and usefulness?
> Automatic summarization: https://en.wikipedia.org/wiki/Automatic_summarization
> "Automatic summarization" GH topic: https://github.com/topics/automatic-summarization
Though now archived,
> Microsoft/nlp-recipes lists current NLP tasks that would be helpful for a docs bot: https://github.com/microsoft/nlp-recipes#content
NLP Tasks: Text Classification, Named Entity Recognition, Text Summarization, Entailment, Question Answering, Sentence Similarity, Embeddings, Sentiment Analysis, Model Explainability, and Auto-Annotatiom
On further review, there are more GitHub projects labeled with https://github.com/topics/text-summarization than "automatic-summarization"; e.g. awesome-text-summarization: https://github.com/icoxfog417/awesome-text-summarization and https://github.com/luopeixiang/awesome-text-summarization , which links to what look like relatively current benchmarks for SOTA performance in text summarization from the gh source repo of https://nlpprogress.com/ : https://github.com/sebastianruder/NLP-progress/blob/master/e...
Show HN: I turned my microeconomics textbook into a chatbot with GPT-3
I never really read my micro-econ textbook.
Looking up concepts from the book with Google yields SEO-y results.
So I used GPT-3 to make a custom chatbot I can query at any time.
How do the responses compare to auto-summarization in terms of Big E notation and usefulness?
> How do the responses compare to auto-summarization in terms of Big E notation and usefulness?
Automatic summarization: https://en.wikipedia.org/wiki/Automatic_summarization
"Automatic summarization" GH topic: https://github.com/topics/automatic-summarization
Microsoft/nlp-recipes lists current NLP tasks that would be helpful for a docs bot: https://github.com/microsoft/nlp-recipes#content https://github.com/microsoft/nlp-recipes#content
ChatGPT is a bullshit generator but it can still be amazingly useful
Prompt engineering: https://en.wikipedia.org/wiki/Prompt_engineering
/? inurl:awesome prompt engineering "llm" site:github.com https://www.google.com/search?q=inurl%3Aawesome+prompt+engin...
XAI: Explainable Artificial Intelligence & epistomology https://en.wikipedia.org/wiki/Explainable_artificial_intelli... :
> Explainable AI (XAI), or Interpretable AI, or Explainable Machine Learning (XML), [1] is artificial intelligence (AI) in which humans can understand the decisions or predictions made by the AI. [2] It contrasts with the "black box" concept in machine learning where even its designers cannot explain why an AI arrived at a specific decision. [3][4] By refining the mental models of users of AI-powered systems and dismantling their misconceptions, XAI promises to help users perform more effectively. [5] XAI may be an implementation of the social _ right to explanation _. [6] XAI is relevant even if there is no legal right or regulatory requirement. For example, XAI can improve the user experience of a product or service by helping end users trust that the AI is making good decisions. This way the aim of XAI is to explain what has been done, what is done right now, what will be done next and unveil the information the actions are based on. [7] These characteristics make it possible (i) to confirm existing knowledge (ii) to challenge existing knowledge and (iii) to generate new assumptions. [8]
Right to explanation: https://en.wikipedia.org/wiki/Right_to_explanation
(Edit; all human)
/? awesome "explainable ai" https://www.google.com/search?q=awesome+%22explainable+ai%22
- (Many other great resources)
- https://github.com/neomatrix369/awesome-ai-ml-dl/blob/master... :
> Post model-creation analysis, ML interpretation/explainability
/? awesome "explainable ai" "XAI" https://www.google.com/search?q=awesome+%22explainable+ai%22...
I for one rue this commandeering of the word "engineering". No, most activities involving stuff are not "engineering". Especially flailing weakly from a distance at an impenetrable tangle of statistical correlations. What a disservice we are doing ourselves.
A more logged approach with IDK all previous queries in a notebook and their output over time would be more scientific-like and thus closer to "Engineering": https://en.wikipedia.org/wiki/Engineering
> Engineering is the use of scientific principles to design and build machines, structures, and other items, including bridges, tunnels, roads, vehicles, and buildings.[1] The discipline of engineering encompasses a broad range of more specialized fields of engineering, each with a more specific emphasis on particular areas of applied mathematics, applied science, and types of application.
"How and why (NN, LLM, AI,) prompt outputs change over time; with different models, and with different training data" {@type=[schema:Review || schema:ScholarlyArticle], schema:dateModified=}
Query:
(@id, "prompt", XML:lang, w3c:prov-enance)
QueryResult:
(@id, query.id, datetime, api_service_uri, "prompt_ouput")
#aiart folks might knowSh1mmer – An exploit capable of unenrolling enterprise-managed Chromebooks
I wouldn't have a career in IT if I hadn't spent many hours at ages 11 to 15 trying to get round my schools network security. My logon was frequently disabled for misuse and I was even suspended for a couple of days once but I learnt more that way than in any class I've ever taken.
I relate to this. As someone currently in high school, messing around with web proxies and code deployment sights, and web-based IDE's trying to run Dwarf Fortress in my school browser has taught me more about computers and networks then just about anything else. It is painfully easy to get around school filters these days. I've never really messed with unenrollment because you do need enrollment to access the testing websites but I've been trying to get into Developer Mode to get linux apps, but the IT guys must have thought ahead on that one.
Chromebooks don't even have a Terminal for the kids. Vim's great, but VScode with Jupyter Notebook support would make the computers we bought for them into great offline calculators, too.
VSCode on a Chromebook requires VMs and Containers which require "Developer Tools" and "Powerwash"; or the APK repack of VSCodium that you can't even sideload and manually update sometimes (because it's not on the 15-30% cut, and must use their payment solution, app store with static analysis and code signing at upload).
AFAIU, Chromebooks with Family Link and Chromebooks for Education do not have a Terminal, bash, git, VMs (KVM), Containers (Docker/Podman/LXC/LXD/gvisor), third-party repos with regular security updates, or even Python; which isn't really Linux (and Windows, Mac, and Linux do already at present support such STEM for Education use cases).
From https://news.ycombinator.com/item?id=30168491 :
> Is WebVM a potential solution to "JupyterLite doesn't have a bash/zsh shell"? The current pyodide CPython Jupyter kernel takes like ~25s to start at present, and can load Python packages precompiled to WASM or unmodified Python packages with micropip: https://pyodide.org/en/latest/usage/loading-packages.html#lo...
There's also MambaLite, which is part of the emscripten-forge project; along with BinderLite. https://github.com/emscripten-forge/recipes (Edit: Micropip or Mambalite or picomamba or Zig. : "A 116kb WASM of Blink that lets you run x86_64 Linux binaries in the browser" https://news.ycombinator.com/item?id=34376094 )
It looks like there are now tests for VScode in the default Power washable 'penguin' Debian VM that you get with Chromebook Developer Tools; but still the kids are denied VMs and Containers or local accounts (with kid-safe DoH/DoT at lesat) and so they can't run VScode locally on the Chromebooks that we bought for them.
Why do I need "Developer Tools" access to run VScode and containers on a Chromebook; but not on a Windows, Mac or Linux computer? If containers are good enough for our workloads hosted in the cloud, they should be good enough for local coding and calculating in e.g. Python. https://github.com/quobit/awesome-python-in-education#jupyte...
Good point. Wasn't aware of the Family Link restrictions. Will see what can be done here.
Disclaimer: I work on ChromeOS.
VSCode + containers + the powerwash feature would enable kids to STEM.
Are flatpaks out of the question? Used to be "Gnome and Chrome" on ~Gentoo.
Shouldn't the ChromiumOS host be running SELinux, if the ARC support requires extended filesystem attributes for `ls -alz` and `ps -aufxz` to work?
Chromium and Chrome appear to be running unconfined? AppArmor for Firefox worked years ago?
https://www.google.com/search?q=chromium+selinux ; chrome_selinux ?
It seems foolish to have SELinux in a guest VM but not the host.
Task: "Reprovision" the default VMs and Containers after "Powerwash" `rm -rf`s everything
`adb shell pm list packages` and `adb install` a list of APKs and CRXs.
Here's chromebook_ansible: https://github.com/seangreathouse/chromebook-ansible/blob/ma...
Systemd-homed is portable. Still, "Reprovision" the broken userspace for the user.
Local k8s like microshift that does container-selinux like RH / Fedora, with Gnome and Waydroid would be cool to have for the kids.
Podman-desktop (~Docker Desktop) does k8s now.
K8s defaults to blocking containers that run as root now, and there's no mounting thee --privileged docket socket w/ k8s either. Gitea + DroneCI/ACT/ci_runner w/ rootless containers. Gvisor is considered good enough for shared server workloads.
Repo2docker + caching is probably close to "kid proof" or "reproducible".
VScode has "devcontainer.json". Scipy stacks ( https://jupyter-docker-stacks.readthedocs.io/en/latest/using... ) and Kaggle/docker-python (Google) take how many GB to run locally for users < 13 who we don't afford cloud shells with SSH (Colab with SSH, JupyterHub (TLJH w/ k8s),) for either.
Task: Learn automated testing, bash, git, and python (for Q12 K12CS STEM)
>> *Is WebVM a potential solution to "JupyterLite doesn't have a bash/zsh shell"?"
"ENH: Terminal and Shell: BusyBox, bash/zsh, git; WebVM," https://github.com/jupyterlite/jupyterlite/issues/949
I actually use a Web Assembly port of VIM on my school computer.
Nice. TIL about vim.wasm: https://github.com/rhysd/vim.wasm
Jupyter Notebook and Jupyter Lab have a web terminal that's good enough to do SSH and Vim. Mosh Mobile Shell is more resilient to internet connection failure.
Again though, Running everything in application-sandboxed WASM all as the current user is a security regression from the workload isolation features built into VMs and Containers (which Windows, Mac, and Linux computers support in the interests of STEM education and portable component reuse).
When Will Fusion Energy Light Our Homes?
It’s lighting my home right now. Looks like clear skies today.
We get free EM radiation from the free nuclear fusion reaction at the center of our solar system; and all of the other creatures find that sufficient for survival.
None of the other creatures require energy to heat their homes, grow food, travel, keep the lights on, or ship cargo in a large inter-connected grid of supply that feeds billions of people.
Everything is also manufactured out of petroleum derivatives. Without it, we go back to making literally everything out of wood and metal, or not making it. 90% of the items you have contact with everyday is made with some kind of petroleum derivative.
EV vehicles are impossible to manufacture without petroleum, so this is certainly a lot more nuanced than just free energy from space...
Given the petroleum is made from creatures[0] which were entirely powered by sunlight, it should be clear that petroleum can be produced by sunlight.
[0] mostly plants, IIRC
Well, not all plastics are made from petroleum and not all that are made from petroleum need to be made from petroleum, it’s just cheaper because there’s a ton left over. Without a petroleum based economy we would either use petroleum for plastics or use other sources of monomers. Plastic would certainly be a lot more expensive, though, especially during a transition phase to other monomer sources. But it’s simply untrue to say petroleum is required to build EVs, it’s just that today that’s what is used as the basis monomers for the plastics.
It’s also factually true, despite a seeming loss of recognition in the last decade, that fossil fuels are not limitless. We might as well not wait for exhaustion before planning about what to do next, no?
Those are "biopolymers", "biocomposites", but like Soy Dream not "bioplastics"?
FWIU, algae, cellulose, and flax or hemp are strong candidates for sustainable eco-friendly products and packaging.
No I was thinking of monomers like bio plastics. But sure, biopolymers etc can work too as replacements. Silicon based monomer/polymers are also very promising. Carbon is a small amount of the earth, while silicon is extraordinarily plentiful. Silicone rubbers are probably the most common example. They use silicon instead of carbon for the backbone.
uh huh, uh huh, and WHY do you suppose EVERYTHING involves the use of petroleum?
It's a classic capitalist evasion of externalities.
Every single drill site is toxic waste disaster, that no one ever has to pay for (except of course, the impoverised who live down stream).
To ignore the many gigawatts of free energy beamed in from space because petroleum will still be used in some way is evading the question.
The fact remains: free fusion power is heating and lighting our homes every time we open a window shade.
And if the combined mafia influence of the contruction-mafia and the petro-mafia didn't have us building houses with NO relation to how they point at the sun, we would use that energy to MUCH greater efficiency...
have you ever been to a drill site? tons where I live. I'd have a picnic out there today if it weren't so cold. Gotta watch for h2s I suppose.
Yes, in fact I worked on them. Enjoy your picnic on top of all the sludge buried next to the wellhead...
Fusion energy is 30 years away and always will be. https://www.youtube.com/watch?v=JurplDfPi3U
This was before ignition was confirmed in a lab setting though so now the countdown has actually begun
"Ignition" was as much a marketing term as anything. It relied on a very specific accounting of energy-in vs energy-out, which no one doubted could be done.
The fact is that they used 50 kWh of energy and produced 0.7 kWh of energy. The fact that at some tiny part of the flow diagram we achieved > 1:1 energy doesn't change the fact that the actual fraction of energy out compared to energy in has barely changed in ten years.
The latest experiment produced a Q-total of 0.014, while before it was something like 0.012.
We can't just hand-wave away the energy in.
Watched the Helion Learn Engineering video, too: https://www.youtube.com/watch?v=_bDXXWQxK38
Has that net-positive finding been reproduced yet in any other Tokamoks?
How does this compare to Helion's (non-Tokamok, non-Stellerator fusion plasma confinement reactor) published stats for the Trenta and Polaris products?
Could SYLOS or other CPA Chirped Pulse Amplification lasers be useful for this problem with or without the high heat of a preexisting plasma reaction to laser pulse next to? https://www.google.com/search?q=nuclear+waste+cpa
Tell HN: GitHub will delete your private repo if you lose access to the original
I was surprised to see this in my email today:
> Your private repository baobabKoodaa/laaketutka-scripts (forked from futurice/how-to-get-healthy) has been deleted because you are no longer a collaborator on futurice/how-to-get-healthy.
That was an MIT-licensed open source project I worked on years ago. We published the source code for everyone to use, so I certainly did not expect to lose access to it just because someone at my previous company has been doing spring cleaning at GitHub! I had a 100% legal fork of the project, and now it's gone... why?
Turns out I don't even have a local copy of it anymore, so this actually caused me data loss. I'm fine with losing access to this particular codebase, I'm not using HN as customer support to regain access. I just wanted everyone to be aware that GitHub does this.
I have a number of git repos that the original developers deleted - because I sync’d them to a usb stick with gitea. I think that is how you have to do it - never entrust a service, especially a free one, with your only copy of anything you value.
If the YouTube algorithm nukes your account and all your videos, you should be ready to upload them to a new account. Same with anything else digital.
My current is standard is one copy in AWS S3 which is super reliable but too pricy for daily use, and one copy in Cloudflare R2 or Backblaze B2 which might or might not be reliable (time will tell) but is hella cheap for daily use.
> because I sync’d them to a usb stick with gitea
Just a tip: no need to use gitea if you want to replicate a git repository to somewhere else on disk/other disk.
Just do something like this:
mkdir /media/run/usb-drive/my-backup-repo
(cd /media/run/usb-drive/my-backup-repo && git init)
git remote add backup /media/run/usb-drive/my-backup-repo
git push backup master
And now you have a new repository at /media/run/usb-drive/my-backup-repo with a master branch :) It's just a normal git repository, that you also can push to over just the filesystemEven better with
cd /media/run/usb-drive/my-backup-repo && git init --bare
Bare repositories don't have a working directory. You can still git clone / git pull from them to get the contents. You can also git push to them without clobbering any "local changes" (there aren't any).More detail here:
https://www.atlassian.com/git/tutorials/setting-up-a-reposit...
Yeah, better in terms of saving space, but I think it confuses some people, hence I didn't use it in my above example. Previous time I recommended a co-worker to use the `push to a directory` way of copying a git repository, I made them create a bare repository, and they ended up going into the directory to verify it worked and not seeing what they expected. Cue me having to explain the difference between a normal repository and a bare one. It also confused them into thinking that a bare repository isn't just another git repository but a "special" one you can sync to, while the normal one you couldn't.
So in the end, simple is simple :) Unless you're creating remote repositories at scale, you probably won't notice a difference in storage usage.
I hear all that, but --bare is necessary in this case because git (by default) won't let you push to a non-bare filesystem branch:
~/temp/a:master $ git push backup
Enumerating objects: 3, done.
Counting objects: 100% (3/3), done.
Writing objects: 100% (3/3), 212 bytes | 212.00 KiB/s, done.
Total 3 (delta 0), reused 0 (delta 0), pack-reused 0
remote: error: refusing to update checked out branch: refs/heads/master
remote: error: By default, updating the current branch in a non-bare repository
remote: is denied, because it will make the index and work tree inconsistent
remote: with what you pushed, and will require 'git reset --hard' to match
remote: the work tree to HEAD.
...
To ../b
! [remote rejected] master -> master (branch is currently checked out)
error: failed to push some refs to '../b'
git clone --mirror
git clone --bare
git push --mirror
git push --all
"Is `git push --mirror` sufficient for backing up my repository?"
https://stackoverflow.com/questions/3333102/is-git-push-mirr... :> So it's usually best to use --mirror for one time copies, and just use normal push (maybe with --all) for normal uses.
git push: https://git-scm.com/docs/git-push
git clone: https://git-scm.com/docs/git-clone
The Qubit Game (2022)
"World Quantum Day: Meet our researchers and play The Qubit Game" https://blog.google/technology/research/world-quantum-day-me... :
> In celebration of World Quantum Day, the Google Quantum AI team wanted to try a different way to introduce people to the world of quantum computing. So we teamed up with Doublespeak Games to make The Qubit Game – a playful journey to building a quantum computer, one qubit at a time, while solving challenges that quantum engineers face in their daily work. If you succeed, you’ll discover new upgrades for your in-game quantum computer, complete big research projects, and hopefully become a little more curious about how we’re building quantum computers.
Additional Q12 (K12 QIS Quantum Information Science) ideas?:
- Exercise: Port QuantumQ quantum puzzle game exercises to a quantum circuit modeling and simulation library like Cirq (SymPy) or qiskit or tequila: https://github.com/ray-pH/quantumQ
- Exercise: Model fair random coin flips with qubit basis encoding in a quantum circuit simulator in a notebook
- Exercise: Model fair (uniformly distributed) 6-sided die rolls with basis state embedding or amplitude embedding or better (in a quantum circuit simulator in a notebook)
- QIS K-12 Framework (for K12 STEM, HS Computer Science, HS Physics) https://q12education.org/learning-materials-framework
- tequilahub/tequila-tutorials: https://github.com/tequilahub/tequila-tutorials
Calculators now emulated at Internet Archive
MAME: https://en.wikipedia.org/wiki/MAME
"TI-83 Plus Calculator Emulation" https://archive.org/details/ti83p-calculator
TI-83 series: https://en.wikipedia.org/wiki/TI-83_series :
> Symbolic manipulation (differentiation, algebra) is not built into the TI-83 Plus. It can be programmed using a language called TI-BASIC, which is similar to the BASIC computer language. Programming may also be done in TI Assembly, made up of Z80 assembly and a collection of TI provided system calls. Assembly programs run much faster, but are more difficult to write. Thus, the writing of Assembly programs is often done on a computer.
I had a TI-83 Plus in middle school, and then bought a TI-83 Plus Silver edition for high school. The TI-83 Plus was the best calculator allowed for use by the program back then. FWIU these days it's the TI-84 Plus, which has USB but no CAS Computer Algebra System.
The JupyterLite build of JupyterLab - and https://NumPy.org/ - include the SymPy CAS Computer Algebra System and a number of other libraries; and there's an `assert` statement in Python; but you'd need to build your own JupyterLab WASM bundle to host as static HTML if you want to include something controversial like pytest-hypothesis. https://jupyterlite.rtfd.io/
Better than a TI-83 Plus emulator? Install MambaForge in a container to get the `conda` and `mamba` package managers (and LLVM-optimized CPython on Win, Mac, Lin) and then `mamba install -y jupyterlab tabulate pandas matplotlib sympy`; or login to e.g. Google Colab, Cocalc, or https://Kaggle.com/learn ( https://GitHub.com/Kaggle/docker-python ) .
To install packages every time a notebook runs:
!python -m pip install # or
%pip install <pkgs>
!conda install -y
!mamba install -y
But NumPy.org, JupyterLite, and Colab, and Kaggle Learn all already have a version of SymPy installed (per their reproducible software version dependency files; requirements.txt, environment.yml (Jupyter REES; repo2docker))Like MAME, which is the emulator for the TI-83 Plus and other calculators hosted by this new Internet Archive project, Emscripten-forge builds WASM (WebAssembly) that runs in an application-sandboxed browser tab as the same user as other browser tab subprocesses.
TI-83 apps:
ACT Math Section app; /? TI-83 ACT app: https://www.google.com/search?q=ti83+act+app
Commodity markets with volatility on your monochrome LCD calculato with no WiFi. SimCity BuildIt has an online commodity marketplace and sims as part of the simulation game. "Category:TI-83&4 series Zilog Z80 games" https://en.wikipedia.org/wiki/Category:TI-83%264_series_Zilo...
Computer Algebra System > Use in education: https://en.wikipedia.org/wiki/Computer_algebra_system#Use_in... :
> CAS-equipped calculators are not permitted on the ACT, the PLAN, and in some classrooms[15] though it may be permitted on all of College Board's calculator-permitted tests, including the SAT, some SAT Subject Tests and the AP Calculus, Chemistry, Physics, and Statistics exams.
Very impressive work. Would be interesting to do some benchmarks versus PyTorch.
On a side-note, I'm not sure if it is because I've looked at so many autograd engines by now, but it is really cool to see that after the years of different frameworks having been developed, most people seem to agree on some concepts and structure on how to implement something like this. It is pretty easy to dive into this, even without being particularly skilled in JS/TS.
Wondering how such frameworks will look in a couple years.
Could there be something like emscripten-forge/requests-wasm-polyfill for PyTorch with WebGPU? https://github.com/emscripten-forge/requests-wasm-polyfill
How does the performance of webgpu-torch compare to compiling PyTorch to WASM with emscripten and WebGPU?
tfjs benchmarks: Environment > backend > {WASM, WebGL, CPU, WebGPU, tflite} https://tensorflow.github.io/tfjs/e2e/benchmarks/local-bench... src: https://github.com/tensorflow/tfjs/tree/master/e2e/benchmark...
tensorflow/tfjs https://github.com/tensorflow/tfjs
tfjs-backend-wasm https://github.com/tensorflow/tfjs/tree/master/tfjs-backend-...
tfjs-backend-webgpu https://github.com/tensorflow/tfjs/tree/master/tfjs-backend-...
([...], tflite-support, tflite-micro)
From facebookresearch/shumai (a JS tensor library) https://github.com/facebookresearch/shumai/issues/122 :
> It doesn't make sense to support anything besides WebGPU at this point. WASM + SIMD is around 15-20x slower on my machine[1]. Although WebGL is more widely supported today, it doesn't have the compute features needed for efficient modern ML (transformers etc) and will likely be a deprecated backend for other frameworks when WebGPU comes online.
tensorflow rust has a struct.Tensor: https://tensorflow.github.io/rust/tensorflow/struct.Tensor.h...
"ONNX Runtime merges WebGPU backend" https://github.com/microsoft/onnxruntime https://news.ycombinator.com/item?id=35696031 ... TIL about wonnx: https://github.com/webonnx/wonnx#in-the-browser-using-webgpu...
microsoft/onnxruntime: https://github.com/microsoft/onnxruntime
Apache/arrow has language-portable Tensors for cpp: https://arrow.apache.org/docs/cpp/api/tensor.html and rust: https://docs.rs/arrow/latest/arrow/tensor/struct.Tensor.html and Python: https://arrow.apache.org/docs/python/api/tables.html#tensors https://arrow.apache.org/docs/python/generated/pyarrow.Tenso...
Fwiw it looks like the llama.cpp Tensor is from ggml, for which there are CUDA and OpenCL implementations (but not yet ROCm, or a WebGPU shim for use with emscripten transpilation to WASM): https://github.com/ggerganov/llama.cpp/blob/master/ggml.h
Are the recommendable ways to cast e.g. arrow Tensors to pytorch/tensorflow?
FWIU, Rust has a better compilation to WASM; and that's probably faster than already-compiled-to-JS/ES TensorFlow + WebGPU.
What's a fair benchmark?
> What's a fair benchmark?
- /? pytorch tensorflow benchmarks webgpu 2023 site:github.com https://www.google.com/search?q=pytorch+tensorflow+benchmark...
- [tfjs benchmarks]
- huggingface/transformers:src/transformers/benchmark https://github.com/huggingface/transformers/tree/main/src/tr...