Pessoas da comunidade brasileira de #ProgramaçãoCriativa por muitos anos fizeram encontros sob o nome promovido pela Fundação Processing, os chamados #ProcessingCommunityDay, fizemos encontros em vÔrias cidades e então depois de 2020, com a pandemia do COVID-19, fizemos três eventos nacionais muito inspiradores em 2021, 2022 e 2023 (vide https://compoetica.github.io/links/)
Ano passado não conseguimos fazer e este ano pretendemos retomar, só que usando outro nome: #Compoética. Vamos aos poucos divulgar mais sobre o encontro brasileiro de programação criativa em https://compoetica.github.io/CP2025/
Meus agradecimentos profundos ao @guilhermesv@guilhermesv que dedica generosamente um enorme esforço para organizar esses eventos da comunidade e cria o design e peças de comunicação sempre emocionantes de lindos.
I hear you, @movq@www.uninformativ.de! :ā-(
At work, too. For a few weeks now when I try to log into this horrible Outlook web intershit (Because why would they fix the Evolution integration?! Itās cactus for well over a year now. Probably more like two.), it forwards me to the corporate weblogin, I enter my credentials, even do the bloody MFA crap and get redirected back to Outlook. āLoading mailboxā¦ā āPlease wait for us to log you out, do not close this window while this process is underway.ā Fuck you! I have to delete the cookies for this damn domain each and every fucking time. Otherwise, this goes in circles forever. I tried the game for 15 minutes, no joke.
But wait, thereās more! Why just fuck it up only a little bit? This week I get logged out at the middle of the day. Every. Single. Day. Not even close to eight hours since I started, no. What the hell!? I reckon I just donāt even bother reauthenticating anymore in the arvo. No more e-mails for Lyse after lunch. Fuck it. Itās just distraction, anyway, right?!
#LGM #LibreGraphicsMeeting 2025 - āThe State of #Processing: How Weāre Bringing a Creative Coding Icon Back to Lifeā
https://www.youtube.com/watch?v=ngQwedwFyOY
#Processing4 #CreativeCoding @processing@processing @processingorg@processingorg
Updating my āhow install and use #py5ā pages, check them out if you want to ā⦠draw and experiment some #CreativeCoding with #Python ā¦ā
EN: https://abav.lugaralgum.com/como-instalar-py5/index-EN.html
ES: https://abav.lugaralgum.com/como-instalar-py5/index-ES.html
Thumbnail novo para a minha pÔgina sobre compreensão de listas⦠#Python
https://abav.lugaralgum.com/material-aulas/Processing-Python-py5/comprehension.html
(preciso dar uma melhoradinha na pƔgina, por umas imagens, arrumar links quebrados)
@movq@www.uninformativ.de I had to look it up! āIs decaf coffee real coffee?ā
āYes, decaf coffee is real coffee. Itās made from the same coffee beans as regular coffee, but the caffeine content is significantly reduced through a decaffeination process. This process involves removing 97% or more of the caffeine, leaving behind the coffeeās flavors and aromas.ā
OK then! š
Fuck 𤣠Building and learning about machine learning and evolutionary processes is hard⢠š¤£
One of the nicest things about Go is the language itself, comparing Go to other popular languages in terms of the complexity to learn to be proficient in:
- Go:
25
keywords (Stack Overflow); CSP-style concurrency (goroutines & channels)
- Python 2:
30
keywords (TutorialsPoint); GIL-bound threads & multiprocessing (Wikipedia)
- Python 3:
35
keywords (Initial Commit); GIL-bound threads,asyncio
& multiprocessing (Wikipedia, DEV Community)
- Java:
50
keywords (Stack Overflow); threads +java.util.concurrent
(Wikipedia)
- C++:
82
keywords (Stack Overflow);std::thread
, atomics & futures (en.cppreference.com)
- JavaScript:
38
keywords (Stack Overflow); single-threaded event loop &async/await
, Web Workers (Wikipedia)
- Ruby:
42
keywords (Stack Overflow); GIL-bound threads (MRI), fibers & processes (Wikipedia)
@bender@twtxt.net @prologic@twtxt.net Jokes aside, I donāt think thatās the right approach either. We had spell checkers, since I can remember, as well as other tools, like the smart image select, used mostly to remove backgrounds. These are tools, that just simplify the process of either opening up a dictionary and looking up a word, you canāt remember the spelling of, or the process of placing a billion little dots around the part of an image you want to select - none of these are creative or enjoyable tasks, we already had tools for them, decades before AI. I donāt think we need to go back to cave paintings, to be free of AIs influence on our creative work.
@prologic@twtxt.net @movq@www.uninformativ.de I donāt even think the premise of this makes much sense. If an artist is convinced they cannot compete, with the āAIā learning models, we already have today, they must have some self esteem issues, strange opinion on what the purpose of art is, or just be someone mindlessly redrawing already established things and not be all that good at it.
It might be connected to some typically non-artists assumption, that the more time and effort the artwork took to accomplish, the more artistic it is - this can be further twisted in these peoples minds, into the āmore pointless detail = more artistic artā meme. AI often ads pointless and illogical details everywhere, āso itās obviously better, than the human artist, who drew the originalā.
Some people just enjoy having the picture they wanted or having the status of an artist to brag about and donāt actually enjoy the artistic process of discovery and small decisions, made while drawing, that shape the outcome into something, only you could have created.
@prologic@twtxt.net Thatās an interesting premise in that article:
The fun has been sucked out of the process of creation because nothing I make organically can compete with what AI already producesāor soon will.
This is like saying itās pointless to make music yourself because some professional player/audio engineer does a better job. Really, thereās always someone or something thatās better than you at a particular job.
If we focus too much on ācompetitionā, then yes, you can just stop doing anything. I donāt know how common this mindset is, especially among artists or creative people. š¤ I would have assumed that many writers, for example, simply enjoy the process of writing. Am I being too naive once more? š¤£
@kat@yarn.girlonthemoon.xyz I only listened to you while going through my photos, so I did not pay very close attention. :-)
Since you have a proper server ā haha, not just one ā and hence are not limited, I suggest you learn a real programming language and donāt waste your time with this PHP mess. It might have improved a wee bit since I was a kid, but it felt like some hacked together shit. The defaults also were questionable at best, it was easier to hold it wrong than right. This stands testament to bad design and is especially terrible from a security point of view.
Youāre right, programming is like any other craft. You only truly learn by actually doing it. And this just takes time. Very long time to master it. Or as close to as it gets. The more you know, the more you realize what else you donāt know (yet). Itās a never ending process. So, take it easy, donāt get discouraged, happy hacking and enjoy the endeavor! :-)
@lyse@lyse.isobeef.org oooh thatās a good point! woodworking is scary and i donāt have much room for it but i do have SOME room in mind that could work for it⦠i feel like iād just hurt myself in the process though LOL
@@twtxt.net The fact that it has an SDK and process management is quite amazing g! š¤Æ
Today I added support for Letās Encrypt to eris via DNS-01 challenge. Updated the gcore libdns package I wrote for Caddy, Maddy and now Eris. Add support for yarnās cache to support # type = bot
and optionally # retention = N
so that feeds like @tiktok@feeds.twtxt.net work like they did before, and⦠Updated some internal metrics in yarnd
to be IMO ābetterā, with queue depth, queue time and last processing time for feeds.
@kat@yarn.girlonthemoon.xyz pandoc is a joy! I havenāt used any Microsoft word processing tools since forever. They want a Word document? Pandoc to the rescue!
Thatās an interesting research article about Wallbleed, a memory disclosure vulnerability in the Great Firewall of China. They reverse-engineered the buggy DNS query processing code that injects a response if the hostname should be censored: https://gfw.report/publications/ndss25/data/paper/wallbleed.pdf
@javivf@adn.org.es Generally speaking if it has been reviewed, discussed and merged, then we accept it as a standard to the set of specs we support. However we might want to document this process and set some guidelines about this to be clear 𤣠Weāve been fairly lax/lose here and I think thatās okay given teh size of our community š
AI problems, top to bottom:
1: Open AI nerds, believe fine tuning a language model algorithm, will eventually produce an AGI god.
2: Subpar artists and techbros who canāt code, convinced AI image bashing and vibe coding, will help convince the dumber parts of Internet, they are a real deal.
3: Parasites, using AI to scam people, because they just want passive income, selling crap, made by an automated process.
Side: Adobe&co, killing Flash/old web, pricing new artists and developers out, to face learning curves of free tools, or use AI, peddled as solution.
twtxt.net
's home page doesn't load more than 13 twts, no more pagination/infinite scrolling...
@aelaraji@aelaraji.com Yeah Iām in the process of rewriting (incrementally) the cache storage backend. Itās now been live for at least a week now and pagination and peering are the last things left to do š¤
The Mastodon admins say that itās probably because of the size of my account (~600 MB), so the export process times out. And I understand that. Here on twtxt, I always use auto-expiring links when I post images or videos. It just gets too much data otherwise. I think Iāll just set my Mastodon account to auto-delete posts after ~180 days or something like that. Nobody cares about old posts anyway.
I saw 100% I/O wait in htop today but couldnāt find a process which actually does I/O. Turns out, I/O wait isnāt what it used to be anymore:
https://lwn.net/Articles/989272/
In my case, it was mpd which triggered this:
https://github.com/MusicPlayerDaemon/MPD/issues/2241
mpd doesnāt actually do anything, it just sits there and waits for events. To my understanding, this is similar to something blocking on read()
. Iām not quite sure yet if displaying this as I/O wait (or āPSI some ioā) is intentional or not ā but it sure is confusing.
@andros@twtxt.andros.dev If something fits in a CSV file, it typically doesnāt require a database. I agree with that. Depending on the application, more complicated queries might benefit from a database, though. I donāt know awk very well, but I could imagine that grep, sed and cut reach their CSV processing limits rather quickly when you have to deal with escaped (multiline) fields.
I only very rarely have to deal with CSV files or databases in my day to day life. Maybe, these classic Unix tools offer some tricks Iām not aware of. When I have some more complicated CSV input, I generally reach for Python.
@prologic@twtxt.net We canāt agree on this idea because that makes things even more complicated than it already is today. The beauty of twtxt is, you put one file on your server, done. One. Not five million. Granted, there might be archive feeds, so it might be already a bit more, but still faaaaaaar less than one file per message.
Also, you would need to host not your own hash files, but everybody elseās as well you follow. Otherwise, what is that supposed to achieve? If people are already following my feed, they know what hashes I have, so this is to no use of them (unless they want to look up a message from an archive feed and donāt process them). But the far more common scenario is that an unknown hash originates from a feed that they have not subscribed to.
Additionally, yarndās URL schema would then also break, because https://twtxt.net/twt/<hash>
now becomes https://twtxt.net/user/prologic/<hash>
, https://twtxt.net/user/bender/<hash>
and so on. To me, that looks like you would only get hashes if they belonged to this particular user. Of course, you could define rules that if there is a /user/
part in the path, then use a different URL, but this complicates things even more.
Sorry, I donāt like that idea.
a few async ideas for later
The editing process needs a lot of consideration and compromises.
From one side, editing and deleting itās necessary IMO. People will do it anyway, and personally I like to edit my texts, so Iād put some effort on make it work.
Should we keep a history of edits? Should we hash every edit to avoid abuse? Should we mark internally a twt as deleted, but keeping the replies?
I think thatās part of a more complete āthreadā extension, although Iād say itās worth to agree on something reflecting the real usage in the wild, along with what people usually do on other platforms.
Added support for uploading images to to #Timeline
Right now you need to copy the markdown code yourself, but next up would be to lean some JS or use HTMX to make the process more smooth.
Also Iām thinking on adding support for If-Modified-Since
since itāll improve the refreshing process š¤
āFirst worldā countries problem number x:
More than 3,600 chemicals approved for food contact in packaging, kitchenware or food processing equipment have been found in humans, new peer-reviewed research has found, highlighting a little-regulated exposure risk to toxic substances.
@falsifian@www.falsifian.org The GDPR does not apply to the processing of data for a purely personal or household activity that is not connected to a professional or commercial activity.
Theyāre in Section 6:
Receiver should adopt UDP GRO. (Something about saving CPU processing UDP packets; Iām a but fuzzy about it.) And they have suggestions for making GRO more useful for QUIC.
Some other receiver-side suggestions: āsending delayed QUICK ACKsā; āusing recvmsg to read multiple UDF packets in a single system callā.
Use multiple threads when receiving large files.
Interesting.. QUIC isnāt very quick over fast internet.
QUIC is expected to be a game-changer in improving web application performance. In this paper, we conduct a systematic examination of QUICās performance over high-speed networks. We find that over fast Internet, the UDP+QUIC+HTTP/3 stack suffers a data rate reduction of up to 45.2% compared to the TCP+TLS+HTTP/2 counterpart. Moreover, the performance gap between QUIC and HTTP/2 grows as the underlying bandwidth increases. We observe this issue on lightweight data transfer clients and major web browsers (Chrome, Edge, Firefox, Opera), on different hosts (desktop, mobile), and over diverse networks (wired broadband, cellular). It affects not only file transfers, but also various applications such as video streaming (up to 9.8% video bitrate reduction) and web browsing. Through rigorous packet trace analysis and kernel- and user-space profiling, we identify the root cause to be high receiver-side processing overhead, in particular, excessive data packets and QUICās user-space ACKs. We make concrete recommendations for mitigating the observed performance issues.
O meu novo salva-vidas na hora de montar um novo site #vuejs sem as tretas dos build systems: Vue3 Tiny Template, da inimitƔvel @b0rk@b0rk
Every time I start a Vue project, I get confused and waste 15 minutes reading the documentation and remembering how to set up Vue.
So this is a tiny template I made for myself so that I can avoid that next time. I donāt use a build process, instead it uses the CDN version of Vue and a single HTML / JS file.
Hurray for password recovery processes.
Started the process of migrating from Github to Codeberg with my projects.
Iām in your node js, messing with your process.env variables.
I remember when doing this process with my wife. During the halfway point we brought all sorts of documentation to show commingling of assets and showing we had ābuilt a life togetherā .. we get to the interview and they just ask if we have a Costco card together. :|
good luck to you!
@abucci@anthony.buc.ci ISO 27001 is basically the same. It means that there is management sign off for a process to improve security is in place. Not that the system is secure. And ITIL is that managment signs off that problems and incidents should have processes defined.
Though its a good mess of words you can throw around while saying āmanagement supports this so X needs to get doneā
who has taken the train to crazytown the furthest? (and hasnāt suffered from nervous breakdown in the process)
sometimes i think i should return to a cleaner state of mind, abandon all big never-to-be-finished projects, and write simple text-processing utilities on a raspberry pi running plan 9, improvising fractile jazz over a lonely lake and spend most of my remaining time meditating.
@screem@yarn.yarnpods.com we have had to really shorten our process. I think long interviews were scaring off talent.
@laz@tt.vltra.plus
How do you handle upgrades like this on your pod? Do you keep a diff of your customisations, or is it all a manual process?
@movq@www.uninformativ.de I am getting this when I run it on cron (extra lines in between becuase otherwise jenny will make them a mash):
Traceback (most recent call last):
File ā/home/quark/jenny/jennyā, line 565, in
if not retrieve_all(config):
File ā/home/quark/jenny/jennyā, line 373, in retrieve_all
refresh_self(config)
File ā/home/quark/jenny/jennyā, line 294, in refresh_self
process_feed(config, config[āself_nickā], config[āself_urlā], content)
File ā/home/quark/jenny/jennyā, line 280, in process_feed
fp.write(mail_body)
File ā/usr/lib/python3.8/encodings/iso8859_15.pyā, line 19, in encode
return codecs.charmap_encode(input,self.errors,encoding_table)[0]
UnicodeEncodeError: ācharmapā codec canāt encode character ā\U0001f4e3ā in position 31: character maps to
@prologic@twtxt.net Excellent, nothing broke. I think what happened was you replied to a twt that I was in the process of editing.
#event Tomorrow, Saturday October 2nd, Iām gonna be hosting a workshop at Processing Community Day CPH about Live Coding Visuals in Improviz. Only 5 spots left, so sign up now at: https://pcdcph.com
When tragedy strikes unexpectedly we cannot just go on as if nothing happened. Our minds need to be given time to deal with the blow. So it is necessary to pause and allow ourselves to process and recover.
@movq@www.uninformativ.de āRandom thought: Would be great if you could do for i in ...; do something "$i" & done ; wait
in a Shell script, but with the Shell only spawning one process per CPU.ā -> Interesting which annoyances stay in the back of the head ā Iād never articulated this, but itās absolutely true that this would be great.
weewiki uses a custom org markup parser written in ANSI C to render the HTML. No emacs needed! my hope is to introduce a user-defined callback that can process these to allow for custom meta-commands.
right now, itās a three step process for me to tweet and upload. gotta make it just one.
I love it. I have a program that needs to processing about half a million records, which will take 3 days. The database that all those records are suppose to go to is acting up after Iāve just done 140K records.
@tx@shroom.party I remember people doing word processing though their spreadsheet program.
I love monitoring a process. Itās more excitement than I can handle.
@mdom@domgoergen.com my own custom client I wrote, I use cron to run the update my timeline every 20 mins. My update process also processes 10 curl calls at time. I did that to save time when I poll everyone.