The world is awash in bullshit. Politicians are unconstrained by facts. Science is conducted by press release. Higher education rewards bullshit over analytic thought. Startup culture elevates bullshit to high art. Advertisers wink conspiratorially and invite us to join them in seeing through all the bullshit — and take advantage of our lowered guard to bombard us with bullshit of the second order. The majority of administrative activity, whether in private business or the public sphere, seems to be little more than a sophisticated exercise in the combinatorial reassembly of bullshit.
We’re sick of it. It’s time to do something, and as educators, one constructive thing we know how to do is to teach people. So, the aim of this course is to help students navigate the bullshit-rich modern environment by identifying bullshit, seeing through it, and combating it with effective analysis and argument.https://www.callingbullshit.org/index.html
This repo contains an annotated overlay for the Nintendo Game Boy DMG-CPU-B chip die and the extracted schematics.
This was done manually with only a few automated checks so THERE’S A HIGH RISK OF ERRORS. I’m in no way responsible if you made someone’s life depend on this and it fucked up.https://github.com/furrtek/DMG-CPU-Inside
This application generates a random medieval city layout of a requested size. The generation method is rather arbitrary, the goal is to produce a nice looking map, not an accurate model of a city. Maybe in the future I’ll use its code as a basis for some game or maybe not.Medieval Fantasy City Generator
“Shepard’s Pi” is one continous song that lasts for 999,999,999 hours, or about 114 years.
Canton Becker’s music generating algorithm composed this music using the first one billion digits of pi (p). Each digit (3.1415…) determines four seconds of music, supplying the “turn signals” used to determine every musical expression.
Because the numbers in pi never repeat, each of the million hours of “Shepard’s Pi” music are in fact unique. By fast forwarding to some distant moment in the song, you are virtually guaranteed to find yourself listening to something that nobody else including Canton himself has ever heard before.Shepard’s Pi
Curtesy of Sam Zeloof I came around the fact that I’ve got a good part of a FSTM in a cupboard here.
Apparently my choice of purchasing the HD-DVD drive for the Xbox 360 will ultimately pay off!! As we all know Bluray won that format war back in the days.
But now it seems that this below would be useable for something:
Over the life of nuclear fuel, inhomogeneous structures develop, negatively impacting thermal properties. New fuels are under development, but require more accurate knowledge of how the properties change to model performance and determine safe operational conditions.
Measurement systems capable of small–scale, pointwise thermal property measurements and low cost are necessary to measure these properties and integrate into hot cells where electronics are likely to fail during fuel investigation. This project develops a cheaper, smaller, and easily replaceable Fluorescent Scanning Thermal Microscope (FSTM) using the blue laser and focusing circuitry from an Xbox HD-DVD player.The Design, Construction, and Thermal Diffusivity Measurements of the Fluorescent Scanning Thermal Microscope (FSTM)
As mentioned, Sam Zeloof shows off the actual chip in more detail:
Here’s one project that aims to develop (and explain along the way) how to create an operating system from scratch. On top of the RISC-V specifics this tutorial also aims to teach how this all can be done in a programming language called Rust.
Keep in mind that all of this is done on a baremetal system. No other software is running.
RISC-V (“risk five”) and the Rust programming language both start with an R, so naturally they fit together. In this blog, we will write an operating system targeting the RISC-V architecture in Rust (mostly). If you have a sane development environment for RISC-V, you can skip the setup parts right to bootloading. Otherwise, it’ll be fairly difficult to get started.
This tutorial will progressively build an operating system from start to something that you can show your friends or parents — if they’re significantly young enough. Since I’m rather new at this I decided to make it a “feature” that each blog post will mature as time goes on. More details will be added and some will be clarified. I look forward to hearing from you!The Adventures of OS
Can you display VGA and play audio on a Cortex-M4 in pure Rust? The short answer is yes, yes you can! Minus the hand-unrolled assembler loop for fixing the phase error in the RGB output. But we don’t talk about that in polite company.Monotron project page
What currently is in place:
- The Atari Joystick interface works, but two Joysticks would be more fun
- The PS/2 Keyboard via the Atmega works, but the pinout was mirrored so you have to put the connector under the PCB :/
- The RTC works
- VGA Output works
- The MIDI Out seems to work when looped to MIDI In, as does the MIDI Though.
- The MIDI In seems to receive data when connected to my electronic drum kit
- The Audio output seems to work quite nicely
- The SD card works, but the power supply can’t handle hot-insertion of the SD card and it makes the TM4C reboot. More capacitors / some current limiting probably required.
I can load games and programs from the SD card into the 24 KiB of free Application RAM. You can interact with these games via the PS/2 Keyboard and Joystick. I can play simple games (like Snake) and play three channels of 8-bit wavetable audio simultaneously. I’ve even got a 6502 Emulator running a copy of 6502 Enhanced BASIC, if you want to go old school!
- Folge mit Miataru HTP001: https://podcast.hack-the-planet.tv/2019/08/09/episode-0/
- Miataru http://miataru.com
- MyFitnessPal https://www.myfitnesspal.com
- New File Menu https://try-catch-finally.net/2019/09/24/easy-file-menu-for-macos/
- Xerox Star https://en.wikipedia.org/wiki/Xerox_Star
- Musashikosugi Typhoon Fotos → https://photos.app.goo.gl/56VtqNNP9unjzQw29
- Emperor Japan Parade https://www.keishicho.metro.tokyo.jp/kotsu/doro/regulation/ceremony_kisei.files/chirashi_e.pdf
- Tokyo Stadtnamen in iOS ersetzen:
- TypeTasker Text Replacement: https://www.schrankmonster.de/2019/09/24/simple-windows-text-expander/
- Cherry G80-3000 https://www.cherry.de/cherry-g80-3000.html
- DasKeyboard: https://www.schrankmonster.de/2012/08/04/daskeyboard/
- DasKeyboard: https://www.daskeyboard.com/
- Kosugi Festa 2019: http://kosugifesta.com/
- Wireless Mac Keyboard von damals: https://geekhack.org/index.php?topic=6528.0
- Synology https://www.synology.com/de-de
- Prusa Mini 3D Drucker – https://www.prusa3d.com/original-prusa-mini/
- OSS Extruder: https://www.sciencedirect.com/science/article/pii/S2468067218300208
- MyFitnesspal selbst bauen: https://www.schrankmonster.de/2019/10/19/replacing-myfitnesspal/
- Public REST API https://www.hack-the-planet.net/2019/09/02/public-rest-apis/
- Swift Programming Language: https://docs.swift.org/swift-book/
- SwiftUI https://developer.apple.com/xcode/swiftui/
- HUD https://www.schrankmonster.de/2019/02/09/head-up-display-esthetics/
- HealthPost https://apps.apple.com/de/app/health-post/id1441893550?l=en
- FHIR https://en.wikipedia.org/wiki/Fast_Healthcare_Interoperability_Resources
- RaspberryPi https://www.raspberrypi.org/forums/viewtopic.php?p=1252786#p1252786
- XNA – XNA’s not acronymed https://en.wikipedia.org/wiki/Microsoft_XNA
- RISC-V https://youtu.be/67KW4t42SZk
- Unicode https://en.wikipedia.org/wiki/Unicode
Back in March 2019 we’d already seen artificial people. Yawn.
Back then a Generative adversarial network (GAN) was used to produce random human faces from scratch. It synthesized human faces out of randomness.
Now take it a step further and input actual small images. Like thumbnails or emojis or else.
And what do you get?
Quite impressive, eh? There’s more after the jump.
Oh and they wrote a paper about it: Progressive Face Super-Resolution via Attention to Facial Landmark
A month ago I wrote about a very black paint. This month brings me a papepr about an even blacker substance.
The synergistically incorporated CNT–metal hierarchical architectures offer record-high broadband optical absorption with excellent electrical and structural properties as well as industrial-scale producibility.Paper: Breakdown of Native Oxide Enables Multifunctional, Free-Form Carbon Nanotube–Metal Hierarchical Architectures
Usually when we visited lectures the notes and explanations where given on a chalk board or a projector. With the lecturer looking away from the audience most of the time.
This is where Light Boards come in handy. They allow the lecturer to face his audience and give explanations on a board…
I was made aways by Ryan Heffernans tweet on the project he did together with his son. He built one of these light boards! The short clip above shows his son on their board.
My son and I built a lightboard. You write on it like a whiteboard, but you can face your audience and the writing is illuminated. Commercial versions cost around $10k, but we made ours for $400 in parts from Home Depot. Here’s how.Ryan Heffernan Tweet
Of course there are commercial ready-made ones. But where’s the fun in that?
To remind you of the recently celebrated 42nd mission anniversary of the still active and data transmitting Voyager 2 space craft.
NASA’s Voyager 2 is the second spacecraft to enter interstellar space. On Dec. 10, 2018, the spacecraft joined its twin—Voyager 1—as the only human-made objects to enter the space between the starsVoyager 2 Homepage
And what reminded me of this astonishing achievement.
Think of this: You are flying at >34k miles per hour. You are >18.5 billion miles away from earth. (It’ll take >16 hours at light speed one-way trip from earth to you). And on top, you are still able to send data back to earth at 159 bytes per second.
I’ve written on this topic before here. And as developers venture more into these generative algorithms it’s all that more fun to see even the intermediate results.
Oskar Stålberg writes about his little experiments and bigger libraries on Twitter. The above short demonstration was created by him.
Especially worth a look is the library he made available on GitHub: mxgmn/WaveFunctionCollapse.
Some more context, of questionable helpfulness:
In quantum mechanics, wave function collapse occurs when a wave function—initially in a superposition of several eigenstates—reduces to a single eigenstate due to interaction with the external world. This interaction is called an “observation”. It is the essence of a measurement in quantum mechanics which connects the wave function with classical observables like position and momentum. Collapse is one of two processes by which quantum systems evolve in time; the other is the continuous evolution via the Schrödinger equation. Collapse is a black box for a thermodynamically irreversible interaction with a classical environment. Calculations of quantum decoherence predict apparent wave function collapse when a superposition forms between the quantum system’s states and the environment’s states. Significantly, the combined wave function of the system and environment continue to obey the Schrödinger equation.Wikipedia: WFC
Right. Well. Told you. Here are some nice graphics of this applied to calm you:
Booting a computer does not happen extremely often in most use-cases, yet it’s a field that has not seen as much optimization and development as others had.
The demon core was a spherical 6.2-kilogram (14 lb) subcritical mass of plutonium 89 millimetres (3.5 in) in diameter, that was involved in two criticality accidents, on August 21, 1945 and May 21, 1946.Wikipedia: Demon core
Now you can have fun without the death-risk in the comfort of your home.
Meet the party-core:
If you’re interested in this topic I can recommend a book:
Atomic Accidents: A History of Nuclear Meltdowns and Disasters: From the Ozark Mountains to Fukushima
A “delightfully astute” and “entertaining” history of the mishaps and meltdowns that have marked the path of scientific progress (Kirkus Reviews, starred review).
I am cycling for fun and for the effect it has on my body and well-being. I do about 30km of cycling every day on average.
After my first stationary trainer broke I bought a new one with the capability to measure wattage and also to apply resistance measured by the watt.
After looking at my average speeds, heart-rates and times on the device I was able to build a quite detailed understanding of the broader picture. What effects my power output and what does not. The effects of nutrition and health to what the body will deliver while being asked the exact same power output curve than the last time.
In a nutshell the numbers tell me that I am usually at a mediocre wattage of 150W constant load doing about 40 km/h average. My reserves usually allow me to go for 1-2 hours without a break doing this.
So far so good. Now I’ve found out from more serious cyclers that there’s something like “Functional Threshold Power“. I do regular have tests at the doctors to check for any heart-rate issues.
Reading about this Functional Threshold Power my curiousity is sparked.
How much could I do? Should I even go for measuring it?
Deep convolutional networks have become a popular tool for image generation and restoration. Generally, their excellent performance is imputed to their ability to learn realistic image priors from a large number of example images. In this paper, we show that, on the contrary, the structure of a generator network is sufficient to capture a great deal of low-level image statistics prior to any learning. In order to do so, we show that a randomly-initialized neural network can be used as a handcrafted prior with excellent results in standard inverse problems such as denoising, super-resolution, and inpainting. Furthermore, the same prior can be used to invert deep neural representations to diagnose them, and to restore images based on flash-no flash input pairs.Deep Image Prior Paper
Apart from its diverse applications, our approach highlights the inductive bias captured by standard generator network architectures. It also bridges the gap between two very popular families of image restoration methods: learning-based methods using deep convolutional networks and learning-free methods based on handcrafted image priors such as self-similarity.
AI and deep-learning is not always necessary or helpful. In this case impressive results have been achieved without the use of any of the hyped technologies.
In this case you give the algorithms two inputs. A video base that you want to stylize and a base picture that resembles the style you want to achieve.
We introduce a new example-based approach to video stylization, with a focus on preserving the visual quality of the style, user controllability and applicability to arbitrary video. Our method gets as input one or more keyframes that the artist chooses to stylize with standard painting tools. It then automatically propagates the stylization to the rest of the sequence. To facilitate this while preserving visual quality, we developed a new type of guidance for state-of-art patch-based synthesis, that can be applied to any type of video content and does not require any additional information besides the video itself and a user-specified mask of the region to be stylized. We further show a temporal blending approach for interpolating style between keyframes that preserves texture coherence, contrast and high frequency details. We evaluate our method on various scenes from real production setting and provide a thorough comparison with prior art.Stylizing Video by Example Paper
Apparently there also is a Windows demo available in which you are supposedly be able to create your own stylized short clips. But as I wanted to try it out it threw a lot of funky messages regarding the application to be specifically untrustworthy / possibly malicious. So be aware and cautious.
When you are writing code the patterns seem to repeat every once in a while. Not only the patterns but also the occasion you are going to apply certain code styles and methods while developing.
To support a developer with this creative work the tedious and repetitious tasks of typing out what is thought can be supported by machine learning.
Chances are your favourite IDE already supports an somehow AI driven code autocomplete feature. And if it does not, read on as there are ways to integrate products like TabNine into any editor you can think of…
Visual Studio IntelliCode is a set of AI-assisted capabilities that improve developer productivity with features like contextual IntelliSense, argument completion, code formatting, and style rule inference.
Of course there are some new contenders to the scene, like TabNine:
TL;DR: TabNine is an autocompleter that helps you write code faster. We’re adding a deep learning model which significantly improves suggestion quality. You can see videos below and you can sign up for it here.TabNine
Deep TabNine requires a lot of computing power: running the model on a laptop would not deliver the low latency that TabNine’s users have come to expect. So we are offering a service that will allow you to use TabNine’s servers for GPU-accelerated autocompletion. It’s called TabNine Cloud, …TabNine
The franchise was started by Kow Yokoyama in the 1980’s. Yokoyama-san was a scratch-build modeller, artist and sculptor. Among his works he built machines of war that would fight in the 29th century, but took their visual cues from early 19th century weaponry and the early NASA space program. All his models were pieced together from numerous kits including armor, aircraft, cars and found objects (like ping pong balls).Ma.K
Often described as WWII machinery in space, these kits are the perfect blend of sci-fi modelling with AFV weathering and finishing techniques…
There is also a longer article with high-resolution pictures here.
With recent announcements around human brain and brain-machine interface research like Neuralink the topic is seemingly seeing some more investments now.
As this whole topic is special to my heart I am interested in all things brain simulations. Thus here’s my personal “logbook entry” on the re-appearance of this topic:
This leads to one of the arguments for whole-brain simulation: it’ll help us solve the “biological imitation game,” a Turing test-like assay that pits digitally reconstructed brains against real ones. Iterations of the test help select increasingly more accurate models for a given task, which eventually become the most promising ideas for how specific biological networks operate. And because these models are based on mathematical equations, they could become the heart of next-generation AI.Singularity Hub
A visualization of my son’s sleep pattern from birth to his first birthday. Crochet border surrounding a double knit body. Each row represents a single day. Each stitch represents 6 minutes of time spent awake or asleepSeung Lee on Twitter
No babies here. But I want such a blanket now.
…in July takes 27.1 days, covering 2967 kilometers. At least if you would have taken the challenge in times of the Roman Empire.
ORBIS: The Stanford Geospatial Network Model of the Roman World reconstructs the time cost and financial expense associated with a wide range of different types of travel in antiquity. The model is based on a simplified version of the giant network of cities, roads, rivers and sea lanes that framed movement across the Roman Empire. It broadly reflects conditions around 200 CE but also covers a few sites and roads created in late antiquity.ORBIS: The Stanford Geospatial Network Model of the Roman World
By simulating movement along the principal routes of the Roman road network, the main navigable rivers, and hundreds of sea routes in the Mediterranean, Black Sea and coastal Atlantic, this interactive model reconstructs the duration and financial cost of travel in antiquity.
I am following the proceedings of ACM SIGGRAPH conferences for more than 20 years now and with the recent years development in computational capacity it seems that many more algorithms and ideas make it to an application near you.
Take this one contribution by Yuanming Hu for example – the Taichi open source computer graphics library:
Taichi is an open-source computer graphics library that aims to provide easy-to-use infrastructures for computer graphics R&D. It’s written in C++14 and wrapped friendly with Python.
Yuanming Hu has been working on the development of Taichi since his third undergrad year (2016), mainly in his spare time. He would like to thank Prof. Toshiya Hachisuka and Prof. Seiichi Koshizukafor making possible his internship at UTokyo, where the initial parts of Taichi were developed.Taichi: About
Also interesting, trivia about the name:
If you ever want to quickly explain what augmented reality could be to a person not knowing yet, you might want to use this (and other) use cases for a visual explanation:
I achieved this by separating the artwork and text into many individual layers, that I placed in receding layers of 3D depth, in a 3D program on the computer. And made sure everything outside the borders of the book is excluded, to give it the ‘portal’ effect.Augmented Reality Book Cover by Alexander Wand
The Chaos Communication Camp is an international, five-day open-air event for hackers and associated life-forms. It provides a relaxed atmosphere for free exchange of technical, social, and political ideas. The Camp has everything you need: power, internet, food and fun. Bring your tent and participate!CCCamp 2019 Wiki
It has been 2005 that I had the time and chance to attend an international open-air meeting of normal people. Of course I am talking about the 2005 What-the-hack I wrote about back then.
This year it’s time again for the Chaos Communication Camp in Germany. Sadly still I won’t be attending. Clearly that needs to change with one of the next iterations. With the CCC events becoming highly valuable also for families maybe it’s a chance in the future to meet up with old and valued friends (wink-wink Andreas Heil).
The Chaos Communication Camp (also known as CCCamp) is an international meeting of hackers that takes place every four years, organized by the Chaos Computer Club (CCC). So far all CCCamps have been held near Berlin, Germany.
The camp is an event for providing information about technical and societal issues, such as privacy, freedom of information and data security. Hosted speeches are held in big tents and conducted in English as well as German. Each participant may pitch a tent and connect to a fast internet connection and power.
Enjoy the intro-movie that has just been made available to us, alongside the whole design material: