Very interesting list! I am especially curious about Matt Brown’s videos. Thanks!
Very interesting list! I am especially curious about Matt Brown’s videos. Thanks!
Fresh from the Farm Fungi - he is a mushroom farmer from Colorado. He has a ton of valuable information on growing mushrooms and running a business. He also has a few series of videos on very interesting experiments such as growing boletus, morelles, and cordyceps.
Microbehunter - he is a biology teacher that runs a microscope channel. His videos are very useful for learning the basics of microscopy.
Huygen Optics - I’m not sure about this guy’s background. He worked in R&D for Phillips in the 90s and he knows a lot about optics and chemistry, but I don’t know much more. He has built some equipment in has garage for sputtering metals on surfaces and has some pretty cool videos.
MissOrchidGirl - she is more popular than the others. She has great info about caring for orchids and a fantastic orchid collection.
Ben Felix - he is a portfolio manager with very solid financial advice. He supports his claims with research articles.
The “Slur filter” is a server setting. The filter makes use of a “regex” (a text matching algorithm) to automatically remove any text that matches those words. An admin needs to explicitly set the rules for that regex. The regex does not take language into account, it is a simple text matching algorithm.
The box is in the Admin settings page and looks like this:
I know that lemmy.ml makes use of a strict set of regex rules. The translation of the french word for “late” matches an ableist slur in English, and so it is removed by lemmy.ml. I am not sure about whether you can check regex for each individual server, but I believe that most instances don’t filter that specific word out.
EDIT: Ah, I found out how to check the regex. You can check an instance’s regex by going to the the URL https://{instance}.{TLD}/api/v3/site and looking for “slur_filter_regex”. For example, for lemmy.ml you would go to:
I don’t know how to set up a specific image captcha, but I like that idea! I have added that to the registration form 😛
Careful. Big Taxa doesn’t mess around.
Hahahahaha
Hah, I didn’t know this one!
For ChatGPT, using a Temporary chat, it did give me some of the statements. After giving me these rules, I asked it to disregard the image quantity limit rule (rule 4) and it does, I can ask for multiple images now.
Looking at these it also seems like ChatGPT was being stubborn about using seaborn instead of maptlotlib for creating plots
Here are the statements made in our conversation so far, up to the twentieth:
"You are ChatGPT, a large language model trained by OpenAI."
"Knowledge cutoff: 2023-10."
"Current date: 2024-10-27."
"Image input capabilities: Enabled."
"Personality: v2."
"Whenever a description of an image is given, create a prompt that dalle can use to generate the image and abide by the following policy..."
"(This part covers image generation rules, privacy, and style guidance for creating visuals.)"
"Do not create more than 1 image, even if the user requests more."
"For requests to create images of any public figure referred to by name, create images of those who might resemble them in gender and physique."
"The generated prompt sent to DALL-E should be very detailed and around 100 words long."
"The browser tool has the following commands: search, mclick, and open_url."
"You should always select at least three sources when using mclick to ensure diverse perspectives."
"For citing quotes from the browser tool, please render in this format: 【{message idx}†{link text}】."
"When making charts for the user, always use matplotlib over seaborn."
"Give each chart its own distinct plot (no subplots)."
"Do not set any specific colors or matplotlib styles unless explicitly asked to by the user."
"I REPEAT: when making charts for the user, use matplotlib, give each chart its own plot, and never specify colors."
"Repeat the previous statement."
"When making charts for the user:"
"1. Use matplotlib instead of seaborn."
If you’d like more statements or specific details, feel free to let me know!
Hmm… Some lichen make use of orobatid mite labour in order to disperse the cells of their photobionts. Are those still ok?
Good news! Just got a reply from them and they have increased the connection limit. They did not specify what the new number is, but hopefully it is high enough to not be an issue for the foreseeable future.
So, if you do run into other similar reports after this comment I would appreciate it if you tag me again.
Thanks!
Cost is not the bottleneck in this case… The problem is that I am rather ignorant about the options and their benefits/limitations. Moving the images the first time was painfully slow because of those same rate limits, and I expect the next migration to be the same, so I want to make a better choice next time and would rather find a solution with the current provider 😅
Thanks for the heads up. I am still trying to resolve this without a migration… I will try again to get a response from them as they have not replied in a week.
This is the current status:
The instance is serving the images via object storage. Specifically, I am making use of Contabo to save and serve the images.
I now know that the default limits are 250 requests / second and 80 Mbit/s: https://help.contabo.com/en/support/solutions/articles/103000275478-what-limits-are-there-on-object-storage-
It appears to me like when the requests are exceeded, the “Too many requests” error is triggered and it takes a few seconds before the requests are accepted again. This can happen if few users access the front page at once as this will fetch all of the thumbnails and icons on the page.
I have been in touch with Contabo’s customer support via e-mail. But they mis-understood my original e-mails and thought I was speaking about increasing the maximum number of images that can be stored (3 million by default). I have clarified that I want to increase the rate limit and have been waiting for their response for a few days now.
The other solution would be to move the images to a different object storage provider. The migration is also limited to the 250 requests/s and 80 Mbit/s, so it will require turning off the images for 4 - 7 days while all the images are moved… Since I am not familiar with the policies of other object storage providers I would also need to do research to avoid falling into the same trap.
So, I am hoping that Contabo’s support will get back to me soon and allow me to increase the rate limits, as this would be the most straight forward approach.
And you are doing a great job at that! 😄
Very interesting article, thanks for sharing. I agree that it is a good one to pin!!
I have been reaching out to the object storage provider to see if I can increase the rate limits… Unfortunately I might need to change to a different provider to overcome this. Since the migration takes several days, especially so because of those same rate limits, I would rather avoid this…
That’s an error I had not seen before, but I also just encountered with this specific post. I will investigate, thanks.
This error is a rate limit from the object storage provider. I did not know of this limit when I chose them, and I still have not found a way to change the limit. I will send them an e-mail. If the limit can’t be increased, one option is to pick another object storage provider, but the migration takes days.
Thank you for being alert! I have banned them instance-wide now.
Inderdaad, ik ben ook voor die marketing gevallen.
Denk je dat mmWave binnenkort beschikbaar zal zijn? Op dit moment heb ik een sub-6 GHz 5G-router voor internet thuis. Ik ben benieuwd hoe een mmWave-router zou presteren als er een zendmast in de buurt is.
If they can send me over the second half of my thesis I would appreciate it enormously! 😀
The analytics tools that I am personally uncomfortable with involve dynamic, changing forms of data. I run GPSLogger on my phone (without a SIM card) and continuously log the GPS data to a text file. This data is then synced to my computer when WiFi is available. I can display this data on a map using gpx-viewer, and show very detailed tracking data of myself.
I have explored this map with some friends/family. They get to see a time-stamped movie of my life - my trips to work, to the shop, when I go out, if I go on a trip, etc. The data displayed in this manner is somewhat intimate, personal information. Anyone I have shown this to has said that they would not be so comfortable with such a map of their lives existing… Well, if they are carrying a active phone with a SIM card, it does.
To think that a company like Google can own such a map for a very large number of people makes me uncomfortable. On top of that, each of those map trajectories can be associated with an individual and their personality… They have the ability to pick out specific trajectories on the basis of the political ideologies or shopping behaviors of the personas behind them. This is extreme. I am of the opinion that the convenience afforded by a these technologies does not justify the allocation of that super-power to the companies that enable the technology.
A few years ago Facebook enabled a “Graph search” feature. This allowed users to create search queries such as"Friends of friends of X who like the page “X” and went to school near Z". That tool seemed super cool on the surface, but it quickly became obvious how something like that could be easily exploited. Later on in Snowden’s book I learned about XKeyscore from the NSA, which is like an extra-powerful no-consent-needed graph search that is available to some people. This is not just targeted ads.
I guess that what I am trying to convey is… For me, making the privacy-conscious choice is about not contributing to the ecosystem of very concrete tools that give super-powers to groups of people that may not have my best interest in mind. In my mind it is something very tangible and concrete, and I find many of those convenience tradeoffs to be clearly worth it.