Small suggestion -- it seems like you're using the wrong units (or maybe abbreviations) for your displayed average page sizes. You're using lower-case b to indicate bits, but I suspect you mean upper-case B to indicate bytes? Also, the lower-case k is the correct prefix for 1,000, but lower case m is milli, or 1/000. You want M for mega, which is 1,000,000.
Also if you really want to be precise you should consider whether you're using binary prefixes vs SI prefixes, e.g. kB (10^3 bytes) vs kiB (2^10 bytes). That doesn't matter as much because the error is small for these lower values, but the casing errors definitely do matter. "mb" means millibars to me, not Megabytes!
Huh... I've always described 2^10 bytes as a "kilobyte" (kB) but I've always hated the ambiguity, even if the difference between 2^10 and 10^3 is usually not important. Thanks to this comment, I learned there is an formal set of units which are distinct from their SI counterparts[0].
Nail on the head. They're just weird, at least in the context of English phonetics. That's the reason I discourage their adoption. I'd much rather we just all agree that kilobyte means 2^10 bytes. Or find words that aren't weird.
If I remember correctly, a big motivation for this change was the fact that disk manufacturers intentionally used base-10 definitions so they could advertise larger numbers for disk capacity. But presumably they still do that, and presumably people still often don't notice.
Sure; there are a lot of distinctions that serve to frustrate casual users. And this is hardly limited to computers - the rant that sticks in my memory is that of a family friend being absolutely infuriated that nuts and bolts can be the same size but have different threads.
The difference does matter, though, and matters a lot when you're working with storage at any scale. So people tend to use the right labels just to avoid ambiguity.
> I'd much rather we just all agree that kilobyte means 2^10 bytes.
Kilo means 10^3 though. I use a lot of SI units every day and that's what it always means for every one of them, just like Mega is always 10^6. These prefixes shouldn't have different meanings depending on the unit being used; that breaks SI. The SI prefixes were first adopted in 1795, before computing even existed as a concept, let alone computers existing as actual objects. The overloading of already very-well-established prefixes to mean something different was always a mistake, and can probably partially be blamed on the US's failure to adopt the metric system.
Why should we use 2^10? Bits matter in the small—a machine with a 12-bit word and 1024 words of memory or whatever was popular in the ’70s—but at the scale of gigabytes, individual bits don’t matter anymore, so may as well just use decimal because that’s what our number system is based on. I don’t see any point besides retro nostalgia to use base 2 after you move out of the each bit counts space.
SI prefixes are fine for mass/block storage and network speeds, because there's no particular reason they would fall precisely into buckets of powers of 2. But for CPU cache and system/GPU memory in particular, and maybe even some flash memory, it does continue to make sense to use MiB and GiB, because of the particular way that memory itself is addressed and packaged. Memory very much does fall precisely along power of 2 boundaries.
For example, I recently bought two 32 GiB DIMMs for my computer. I guess you could call them 34.3597 GB DIMMs, but that's strictly worse! Knowing that they're exactly 32 GiB makes it makes it obvious that it takes 2^35 bit pointers to address every location by byte in one of those DIMMs (so they obviously require a 64-bit architecture to take advantage of!), or 2^36 bit pointers to address memory locations across both of them.
Notably, storage vendors and many operating system vendors use different units to describe capacity, so hard disks will always seem small as a result of this.
This is a hilarious nit-pick. In the context of an internet blog describing average page-size, it's completely obvious that mb == MB ... not millibars...
There's an insane amount of abbreviations and acronyms that have multiple meanings in different contexts. How many Wikipedia pages have [disambiguation] here?...
It is not disambiguous, it is wrong. People will probably figure out what it is meant from the context, but it will slow them down. Are there really any reasons for using the wrong abbrevation other than lazyness or ignorance in this case?
Maybe not officially, but in practice that's how it's used:
The unit's official symbol is bar; the earlier symbol b is now deprecated and conflicts with the use of b denoting the unit barn, but it is still encountered, especially as mb (rather than the proper mbar) to denote the millibar.
Yeah, in general, there are many, many "unit collisions" that can really only be exactly interpreted from context. I think it would be great if everyone started using bracket notation (or similar) for prefixes. E.g. [k]B [Mi]B This is the convention used in pqm.js and it works really well. this would go a long way toward units that can be accurately read by a computer.
But is the SI system ambiguous? I almost never have to think about prefix v unit, it is always quite clear. Of course, strange combinations can occur (MNm - Meganewtonmeter as an example perhaps), but even those are unambiguous on second thought.
Collisions seem to be most prevalent in "IT" units or for improper SI usage. But maybe the latter is really the reason for the former in this case: if people separated SI and other units properly for IT units, it seems to be it would also be perfectly fine (but like me, a lot of people seem to have no idea of the correct definitions).
If you are strictly sticking to the SI system of units, you should be fine. However, some of us work in industries and countries (You know where) that don't fully embrace SI, and mixing other systems with SI is common.
Pressure measurements/specifications are part of my daily work. I have never seen "b" referring to bar, also not in any of the many American (imperial) papers I have come across. So this is either properly old usage indeed, or very specific to certain regions or industries (like usage of relative v absolute pressure). Or Wikipedia being properly pedantic.
You should learn to understand MB/mb/Mb/mB for megabyte and mbit for megabit. Anything else is inviting error and sorrow, because many people use it that way.
This is flat-out wrong. Most times when I see a lower-case b in software or documentation it really does mean bit, not byte. The standards exist for a reason and you should follow them. Anything else is inviting error and sorrow.
I mean, if they wanted to reserve the use of this common animal name across all industries, that’d be ridiculous, but it would seem pretty reasonable to worry about confusion between a cloud notes app and a cloud blog app. Certainly easier to think about it now than if this service takes off.
These are two different products. Blogging != note taking. And it doesn't seem like there's malicious intent. You can't use a generic name and expect others not to come up with the same idea.
Blogging !== note-taking, but you could definitely make the argument that a blog post is semantically a public note.
Apple, Fox, Shell, Target...any of these generic words ring a bell?
I’m sure there was no malicious intent, and it’s very possible that these two products will coexist without any further issue.
But as evidenced in this very thread there is the potential for confusion, which is the whole point of trademarks, and the reason why I’d have an easier time incorporating “Apple Surfboards“ than I would “Apple Keyboards”.
I saw the headline and immediately assumed it was an app by the same Bear company. I would guess that almost every single Bear Notes user would have the same reaction.
Is it known among non mac/ios users? I've never heard of it before. Not to mention that by using such a generic name you're just begging for collisions.
I agree, bear is great. I miss it - I no longer have an iPhone and am Windows dependent :-(. I have made the best use of OneNote as I can - but the ease of categorisation of thoughts within the bear app will always be the best imo.
I want to love Joplin (open source! markdown!) but I just can't get past how ugly it is. Its iPhone app is just a horrendous mishmash of colors and non-native UI elements.
Joplin is cool because you can add your own css to customize the desktop app. I realize not every is interested in doing that, but I’ve been an advocate once I tweaked the styles to my liking [0]. Agree with you that the mobile app styles are very rough, but the community is very active and open to suggestions / improvements [1].
I wish Notion had a self-hosted version. I store a lot of things in Joplin (previously Evernote) that I'm just not comfortable having on 3rd party servers :(
I have even stronger feelings about Roam though. Notion is "better" than just abt anything else like it, whereas Roam is different -- there is nothing else like it.
Since HN seems to be on a blog-kick lately, I'll repost the idea that I'm still waiting for someone to build:
A blogverse of some kind that allows for algorithmic discoverability & aggregation (ala Medium) without the bullshit/terrible UX.
The real value proposition of Medium is that a well-made aggregator benefits readers and writers alike. Readers find more authors they like, writers find more audience. There are also network effects with shared comment logins, inter-blog citations, etc.
I really think a blogging renaissance is waiting to happen. These ingredients plus a business model not reliant on ads, massive js overhead, and other nonsense could jumpstart it.
What I want is a Yahoo-style directory for blogs. Blog owners can put their blog in exactly one category. Users can star the blogs they like, similar to GitHub, and identify the low quality click generators/marketing blogs.
I personally find Medium to be a horrible way to find content. Maybe it works for new content, if that's what you're after.
The success of search engines prompted Yahoo to ditch its original product (the curation and categorization of the web) in favor of its competitors automated (and thereby game-able) crawl-index-search approaches.
Now, decades later, there seems to be a shared yearning for the curated web, perhaps in response to the low signal:noise ratio of search. Isn't it funny, how the world works in cycles?
Curated search (domains chosen by a set of humans with no financial conflicts of interests, with some grokkable categorization and full-text search) might be the nirvana we're searching for. I think the GP has a point, that the need for a sustaining business model tends to strongly conflict with this equilibrium.
Wikipedia has sort of evolved to partially fill this niche, but it periodically struggles with funding. I agree there's not a similar filling for blogs, yet: maybe GitHub will evolve there, but it will face the same pressure of other platforms owned by public for-profit companies.
I was probably one of the last users of their original directory. It was my browser homepage until the day they removed it.
I eventually stopped using it because it didn't keep up. These days, search is for many purposes completely useless. If I want to find someone to do work on my house, the last place I'll go is Google. It's truly amazing just how worthless the results are. You'll get results from Michigan and Florida and Oregon all for the same search, in the same town, and claiming to be a local business. I imagine it's a fraud-ridden garbage dump if you actually try to use Google to find businesses to do work for you.
On the topic of blogs - not completely useless, but overrun with shallow, uninformative trash posts by SEO experts. I think Google is more vulnerable now than at any time in the past 20 years.
> You'll get results from Michigan and Florida and Oregon all for the same search, in the same town, and claiming to be a local business.
It's worse if you are non-US. Google seems to think that anywhere in the UK is local to me for businesses, and that's AFTER I've added loads of filters to stop American results from showing.
> Curated search (domains chosen by a set of humans with no financial conflicts of interests,
The problem is that the second it becomes an authoritative source, every spammer and marketer will start trying to game it, just like the SEO does now. Corruption is a huge problem with that approach. I'm old enough to remember how much influence yahoo categories editors had, and that many have asked for money to include you in the list, or they would use the position to simply block all the competition to their own sites. Same story is happening with Wikipedia, just it's more about personal wars and agendas instead of straight down racketeering.
Yeah, this is/was a real problem, and I don’t know if there’s a perfect solution. Anecdotally, @dang and friends do a very good job “editing” hn. They’re presumably paid well for their job, and seem to be passionate about their roles (thank you!).
I feel like the key to good curation might be good moderation. In other words, allow people to submit links to be on the list, but there must be humans to control for SEO and marketing spam (like there is on HN). Making those humans incorruptible is hard but maybe not impossible, if they’re well compensated and hired thoughtfully.
That being said, I’m doubtful that such a strategy of moderation would scale to hundreds of thousands (or tens of millions, shudder) of active voters and submitters. At some point, it would need to federate, with different mods owning different lists, and it would be difficult to avoid a devolution into Reddit.
> Curated search (domains chosen by a set of humans with no financial conflicts of interests, with some grokkable categorization and full-text search) might be the nirvana we're searching for.
Moving away from ads towards human-centric curation is a primary design feature of the decentralized web.
I run a Yahoo-style directory for blogs and articles I find: https://href.cool/. I'd personally like to see more personal directories rather than big monolithic directories of everything - those tend to collect spam.
Well, it’s my opinion that technology can’t solve the discovery problem. I know we want it to. But at some point the technology has to evaluate the content. It can’t - so technology gives the content to humans to evaluate. However, it can’t evaluate the humans’ capabilities. :/
You discover a directory like mine just as you would discover any other link, johntash - by coming across it as you read, perhaps on Hacker News. If there were more directories, they would be easier to discover. They happen to be richer discovery points than a normal blog or profile page.
Fun anecdote -- in the late 90s I maintained several web sites (personal/hobby ones) and by far the best source of traffic for them was Yahoo directory pages. I didn't know the SEO game (if there was even one at the time) and while search engines brought a small amount of traffic, majority of visitors came from Yahoo. When I launched a new hobby site, the goto marketing plan was to apply to add it to Yahoo. Granted, we're talking about the range of dozens to maybe a hundred or two visitors a day, so it's not web scale :)
A question for you (and everyone): how important do you think it is that the algorithmic discoverability be married to a single platform/aggregator? What if you just had better algorithmic discoverability across all writing on the internet, regardless of where it's hosted?
I get that that would miss some of the benefits you mention (shared comment logins for example) but I'm wondering if people think it would capture 80% of the same benefits or, like, <50%. And I don't think for the discoverability to work there's any innate reason it has to be restricted to a single blogging platform.
Personally, I've started seeing algorithmic discoverability as an anti-feature that mainly serves the interests of the attention economy.
For my part, I want more long-form, thoughtful articles that offer an enriching read. Not only have we got a good decade or two's worth of experience showing that no algorithmic discovery system ever favors that kind of thing over content that's morally equivalent to Five Minute Crafts, but I've got a couple decades' worth of experience telling me that, since long-form bloggers tend to link each other quite a lot, never needed an algorithm to help me with discovery in the first place.
If I want anything, it's algorithmic filtering: Take the feeds I'm already subscribed to, and filter out the stuff that I tend to skip over without reading. Because my blog feed already delivers me a couple hours' worth of reading a day, so I can afford to be choosy.
Yeah, this is very true. I may have overstated my initial idea. When I think more about it, maybe what I'm looking for is handcrafted curation, but that let's anyone share their curation speccs and discover relevant things to curate from. To me, that suggests some kind of algorithmic component, maybe in the form of search, voting, etc., but I wouldn't necessarily want the whole thing driven by either machine or mob.
Like stitchfix, but for magazine articles, blog posts and books. I might be cynical, but I think humans do this much better than any algorithm I've ever seen. :)
For myself, I usually read in two "phases". I collect material from sources that are usually high quality for me, and do some superficial skimming to trim out the content I don't particularly care for.
Later, when I'm more "in the mood" or have a long period of time, I pop() and read(). The stack gets pretty large at times, but it only takes a couple days of vacation to blow through most of it. If I pass over something in the stack enough times, it gets free()'d.
I don't know if I'd trust an algorithm to do either of these stages for me. I'd trust an algorithm provide input to phase (1), but never to replace phase (1).
> If I want anything, it's algorithmic filtering: Take the feeds I'm already subscribed to, and filter out the stuff that I tend to skip over without reading. Because my blog feed already delivers me a couple hours' worth of reading a day, so I can afford to be choosy.
"All writing" is probably too broad, but I think there's an interesting "meet-in-the-middle" solution where the site is a kind of syndication machine that is creating traffic between separately hosted blogs.
I appreciate people wanting to own their own writing, having permanence of data assured, etc. so this is probably worth thinking about. At the same time, many outside the tech world don't want the overhead of hosting or setting up their own blog.
I wonder if a hybrid would be best? Host yourself option or host-by-us option, with a pricing structure that accords.
The point is really to maximize ones ability to traverse between writers on subjects that interest you, so the physical location of the data is secondary.
I tend to be skeptical that "$x but better" is a viable business in the VC era or maybe on the internet at all. On the internet, where publishing is easy, but distribution is hard, it seems like there's already a natural tendency toward winner-takes-all. Execution matters somewhat, but at some point, I think it comes down to who can throw more capital at the problem.
I'm interested in this problem space, because current social media tools leave me somewhat dissatisfied. But I'm also skeptical about how you'd build it successfully.
- Bootstrapped with low overheads. No outside investment removes pressure to grow at all costs and the lapse on integrity that we see more of each day.
- No data lock-in. Export your posts in a single JSON file (containing Markdown + HTML versions) accompanied by all images.
- We're finishing up a big set of data portability / data respect features, which we plan to announce soon.
The aim is to build a community of capable people with a genuine interest in technology and attach a job board to the site. Companies can then pay to display their vacancies on the job board and users can take a look whenever they like. No popups, banners or any of that dodgy/spammy crap getting in between users and the reading/writing. In this age we see integrity towards our users as a differentiator.
Building it is fairly straightforward and fun. Some learnings we've gained in terms of 'jumpstarting' it:
- a lot of people are vehemently sceptical after the Medium debacle.
- creating a feed that prioritises good content without "censoring" dev spam or trivial posts is tricky when coming off a smaller user base. For example, upvotes can have outsized effects.
- getting regular volume of good content so that people use Able as source of news/inspiration/learning.
We've had some great posts from people but we need just that little bit more to get the flywheel going. As soon as someone posts something, activity on the site goes up but then dies down again. It's the classic building vs. promoting trade-off. However, we've chosen to get data portability and respect right first as we believe this is fundamental. We're wrapping that up now and then have loads of learnings and ideas we want write about.
We feel that same potential 'renaissance' you're talking about and this is how we want to try and activate it. If you feel inclined, have a look and let us know what you think.
Regarding the upvoting problem: Try using only downvotes and sort by downvotes*age. This should solve a lot of problems that usually come with upvoting systems, e.g. false negatives (good articles with bad score).
Personally, many of my good reading suggestions come from a trusted network. I trust a few people to only recommend content of a type that is high-quality and interesting to me.
I do think it's important to have an element of discoverability for new content (otherwise there's no real way for a writer to bootstrap into visibility), but I think an important element is being able to follow and trust content aggregators. Maybe algorithmic curation will be trustable in a decade, but right now it feels too gameable, and too easy to degenerate into thinkbait.
I also felt that I value personal recommendations much more than from huge aggregators or algorithms. I wanted to give these 1-to-1 recommendations a better vehicle than WhatsApp. So I created an app for it: https://onelink.to/listo
It's still early, but already usable.
Medium has a revenue sharing thing where a medium pro users fees are allocated to write depending on what the user reads. I don't know anyone who has medium pro though.
Sure thing. Going to open source it after it's done. I have been thinking of using activity pub and making discovery decentralized so people can host their own curation and share with each other.
I don't even want to have a medium.com account for their crappy UI and hostile paywalls, but I doubt those who use medium regularly actually discover new blogs.
HN is one of my primary sources to read new things, and I'm already blind to medium's footer read-more links.
* Good clean default CSS so I don't have to mess with it
* Automatically upload website to CDN and trigger expirations as necessary.
* Self-hosted
* Runs on AWS Lambda or any other Function as a Service equivilient
* Has a super lightweight CMS that I can easily use on both desktop and mobile, so if I have ideas I can start writing anywhere, and can also make minor corrections to existing posts while on the go.
* The CMS can be a frontend to git, but git is hard to use on mobile, so I don't want the CMS to just be git.
If anyone knows of something that meets these requirements I'd be super grateful!
Looking at your profile, I get the sense you are trolling :) but I have a bad time imagining the need for the elasticity of lambda for those single-user actions.
I'm not sure whether it's what they meant, but I could see the "admin" cms portion being hosted on lambda.
You'd load a static editor page, write your content, and then post it to a lambda endpoint. That lambda would then generate static pages for your content and push it to s3 or wherever.
We're also considering providing our CMS as headless via an API so that you can connect it to your SSG and make your own template, host it wherever you want, etc, but we haven't decided yet on the pricing for that but it would be much cheaper than our main product.
Sounds like you just need a way to modify a remote filesystem, then trigger a build step when you're done with the modifications.
Assuming it was wrapped up in a nice UI, would that be missing any features you need?
EDIT: Put more concretely, if you had a site hosted on S3, all you need is a way to modify the source files of that site, either directly or by modifying files in another directory and running a transform step to produce the final result?
Sounds like previews is the main thing missing though? I don't consider github to be similar to a remote filesystem, because like you said certain simple things like uploading images and other arbitrary data isn't easy.
There's no simple commercial solution that does exactly this that I know of, but you can get very close with Google Cloud Functions, a GitHub Pages site, and a Google Sheets spreadsheet as a datastore. Not static, but it is fast, and you get a 'CMS' for free (the spreadsheet). Plus all of these services can be used completely for free.
I'd advise just running Wordpress or Ghost. Even a small VPS is more than enough scale and you get a working blog out of the box with easy extensibility for the future. And you can still use a CDN in front.
People think a static site is enough without realizing SEO, RSS feeds, comments, etc, are all things you might need and would have to rebuild yourself.
With self hosting I can control the code, what it does, the cost, where it stores data, how backups are made, when new versions and updates are deployed, etc.
It’s always a trade off on control, and my comfort point is “on AWS”.
For some people they want their own servers. Some people want to own the network block. It’s just what level of comfort you have with each type of control.
I don’t know what CMS I’m going to use, that was my question. Which one operates the way I want?
It's a single-file app. You can host it yourself, and it edits static files in a github/gitlab repo. The deployment after push and hosting with custom domains and CDN is already handled by Netlify. And it's all free.
If you use Jekyl then you can skip Netlify and Github will build and serve your site.
I'm struggling to understand what you mean by self-hosted. What does AWS Lambda do in this scenario? Surely not host the blog, so it must be the "cms"? At that point I'm not clear why you'd rule out netlify, as it's no longer "self-hosted." Please clarify! :)
* this triggers a lambda function that generates a static website and puts it on s3
* this triggers another lambda function that pokes the CDN and expires anything that changed
By self-hosted I mean I control all the parts. I can change the generation code, the CDN upload/expiration code, I can change the output, I can view the logs, all within my own account.
I don’t know... I have a Github actions workflow that on a commit, it builds a site (mkdocs) and upload it to a gh-pages branch for hosting on Github. You could also push the generated html to S3.
That coupled with Working Copy could be a pretty good mobile workflow.
For me, this was much easier than deploying Lambda code and managing How all of that worked together. Using Github Actions wasn’t much more than setting up a deploy script.
Have that. It sucks. UI on github sucks (need something I can access anywhere, not just my desktop, thought github might cut it, it doesn't). Uploading images sucks. No preview sucks. I suppose I could trigger to staging but that means I have to commit just to see a preview then wait for the entire process vs say wordpress which has instant preview.
Managing posts also sucks. I can pre-date posts but I can't at a glance get a list of posts with data about each one. Instead I just get a file listing and have to open each file and read its headers to see its metadata (date, title). Publish vs unpublished also suck. I have folders for my posts and another folder for drafts but moving something from one to the other is no fun compared to just checking a box in wordpress.
I've build something like this using Hugo as a SSG and a custom Micropub-backend that commits new content to a Git repo and pushes it to a Git hosting platform. That then triggers a CI build for the site generation and upload to the webserver as well as a CDN purge.
Netlify CMS uses the GitHub (or bitbucket, etc) API to modify your repository with the new/updated content. You can self-host the entire system, or have netlify handle the GitHub auth while you self host your website.
You could use GitHub actions to do the build and send it off to some CDN to deploy, or if you were self-hosting the whole thing, you could have a Flask app listen for a webhook and re-build your site when that's received. Or simply pull in from the git origin and rebuild with a cron job.
I guess what I'm getting at is doing that would be a big patchwork of putting all that together. Not a unified system. I'd gladly make a one time payment for a unified piece of software that does everything (or of course gladly adopt something open source).
Also, to your specific example, does there exist a Flask app that runs on AWS lambda that can listen for a webhook and then build a static website?
Does there exist code that I can put into Github actions that builds a static website and uploads it to a CDN?
That's basically what I'm looking for.
Of course I could write my own solution or cobble it together with a bunch of moving parts, but my whole point is that I'm looking for an existing package that already does it.
If you use a tool like the Serverless Framework, you can "cobble together the moving parts" and store it into GitHub and make it easily deployable into the cloud. Putting together pieces is what software development is. If you are used to OOP this is even more relevant as the entire idea of OOP is to create discrete pieces of code and compile them together.
PHP for example used to be a "big patchwork" of seperate files working together, but the right tooling came along to think of it as "a unified system". Same thing with Serverless and the Serverless Framework
Shouldn't be too difficult to hook this up to a build command for hugo, jekyll, or whichever static site generator you'd prefer. But this gets away from the self-hosted part of what you were saying a bit.
With self hosting I can control the code, what it does, the cost, where it stores data, how backups are made, etc.
It’s always a trade off on control, and my comfort point is “on AWS”.
For some people they want their own servers. Some people want to own the network block. It’s just what level of comfort you have with each type of control.
What if your content doesn't have enough views / you get bored / life changes so can't afford server costs anymore / die. Your server will expire eventually, and there goes your content. web.archive.org might have some sites archived, but many blogs won't have been archived so their content is just gone forever.
I've self-hosted many platforms, and many have died, perhaps due to running costs or lack of need anymore. It's partially why I'd never self-host my email, for example.
You can't know the future, and if your content is gone, it's gone for all of your audience (or potential future audience). Perhaps an ideal solution is some kind of self-hosted platform that mirrors content into a forever-public external repository, and hence preserves it, even if your hosting ceases to exist.
Yeah. There's a wordpress.com blog I frequently go back to and reference and the author passed away in 2014. Thanks to it being hosted on wordpress.com it will stay up there for the foreseeable future. Had it been relying on monthly payments for hosting and yearly payments for the domain it would probably have been down already by the time I realized he was no longer with us (some three months after the fact). And at that point, not all content would've probably been in archive.org and thus lost forever.
I also have a bunch of old blogs that I've just given up on at some point and now they're gone forever.
Your last point is kind of what archive.org's Wayback Machine does. Haven't checked whether they have an API where you can submit URL's you publish* or if you'd have to submit new posts manually, but this could be a decent solution. It wouldn't cover non-tech people though.
*) I just now quickly browsed through their API doc but couldn't find anything clear, other than uploading random files
I think that product template could use either `aria-hidden="true"` or `aria-label="bear"` HTML attributes for screen readers sake (and reconsider title and OG properties); not an expert in this area nor having SR at hand, but I guess that
ʕ•ᴥ•ʔ
would sound like 'pharyngeal voiced fricative - bullet - letter ain - bullet - glottal stop', what is hardly beneficial for screen reader users. Cool Unicode "picture" though, it's a pity such doodles hurts accessibility (sad smiley).
Given that we can't expect everyone to use aria attributes, shouldn't screen readers just have a list of all widely used smileys with descriptive names (if they don't have it already)?
Good point and interesting question indeed. Finally tried simple speech synthesis demo [0] and in Win10 Firefox it really reads out some "basic ASCII-smileys" as their descriptive translation (and ignores any other set of non-alphabet characters, with few exceptions like underscores and asterisks).
Compiled all predefined textual emoticons offered by Gboard app in Android [1] and let the Narrate speech synthesis in Reader view (F9) read them aloud (uses presumably the same synthesis as example) - quite predictably most of them weren't heard at all, but those simple ones that were have surprised me.
So most probably there is no problem with that particular Unicode bear in screen readers after all.
What are the limits on number of posts, post size, images...? What is it going to be priced at? Will there continue to be a free tier if/when it becomes paid? Sorry, I can’t try any platform that doesn’t answer questions about its future. I hope you’d provide more details on the homepage.
I'm slightly amused by the domain hacks, but also concerned about a few of the ccTLDs, especially when it comes to user generated content and blogging/opinions. For example, the leader of .ph routinely calls opponents gay, is not known for human rights or free speech, and I am curious how this might reflect on it.
I have the same concerns with user generated content in/on .de and their Netzwerkdurchsetzungsgesetz law, the mandatory Posting an Impressum on your sites, etc.
I have seen bitly suspended before by Libya for merely providing redirects to content they don't like.
For anyone who is interested in this - it is great and you can export all your Apple notes.app stuff (inc images) with the mac app Exporter, then import into Bear.app.
I can't recommend Bear enough if you're using all Apple devices. I've used many note taking app but Bear is the only thing that gives me Markdown note taking + not getting in my way of writing. I especially like the infintely nested tag of Bear, such a time saver when you can drop a #hastag/anywhere/with/inifitely/nested/hierachy .
Disclamer: I'm not affliated with Bear in anyway, just a happy customer.
Bear uses iCloud storage which has on multiple occasions trashed people's data during iOS beta periods. It's not "Bear Sync", it's just another iCloud beta sync problem. If you value your data, don't use iOS betas. It's that simple.
If they had the choice, many developers would disable their apps on beta devices because of issues like yours.
I used to use bear too until I discovered notion.so, now I use nothing else (also not affiliated with Notion or Bear, but used to use bear for everything)
I'm surprised nobody has attempted to put the content in the URL yet (to display on a static page with styling using JS [needs a tag filter...] to insert an URL parameter into some node). It would accommodate at least 2KB of text, local caching and fast hosting all in one.
I was always quite fascinated by the concept, but I suspect liability and lack of control over the content is a fatal issue and why nothing much seemed to come from it.
If someone makes a 'bad' page, which is inevitable, the domain with the hashify/shortly code would be held responsible and the only way the site owner could 'remove' the content would be to stop the service.
> Storing a document in a URL is nifty, but not terribly practical. Hashify uses the [bit.ly API][4] to shorten URLs from as many as 30,000 characters to just 20 or so. In essence, bit.ly acts as a document store! [1]
bit.ly et al. seem to be able to get away with being agnostic processors. I'm surprised there haven't been more stories about their services being abused.
Thinking through this, it seems like content-in-URL would work for a website with a single page or a small number of pages, but would limited by the fact that links from one content-in-URL page to another content-in-URL page require content from both pages to be encoded in the first page’s URL. If you have pages linking to pages linking to pages, this cascades into requiring content from all pages encoded in the home page URL.
I guess I'm a bit late to the party, but after reading your comment I hacked something together - https://x.rukin.me It's ugly and I haven't spent any time styling it or improving editor(it's just textarea) but it works as PoC.
Also I've started it before I read comments that hashify exists, otherwise I would probably not do it ;)
Compared to hashify:
- it uses zlib to compress data, so it can actually contain much more content(if it's repetitive)
- it also supports password encryption of the content (don't know why, but my friend said it would be cool)
- only supports markdown(no html) as I haven't found any good js lib for html sanitization on client side
This has been done, iirc. I think it was HN where I saw it a couple years back. I couldn't find it today if I wanted to, but it's definitely been done.
Of those it seems only telegra.ph has a good UI allowing easy link/photos embedding, and the result is really pleasing. All the others need to rely on a third party. Is it open-source ?
I like this idea a lot and the design is wonderful. But, I would have liked to see a link to an example blog on the page just to see what the output looks like.
If I may, I would like to add my own Markblog to the list :-) It is basically a static site generator based on markdown files. (https://github.com/olaven/markblog)
I appreciate the no-bs lightweight website sentiment as much as anyone, but I think there's also something to be said about drastically improving readability with some line-height and font styling.
write.as does support custom CSS and JS, so you can, in theory, make look however you want. It isn't quite as light as Bear is, but it's no Medium either.
oh interesting! Yeah I actually hadn't been to their landing page in a long time, totally agree with your there (and feel like it doesn't do their actual blog product justice - to my tastes at least their blogs are perfect, low-key but elegant and very readable).
It looks simple, but isn't technically minimalistic with 77KB .css and loaded fonts. The first page load was actually visibly slow with the fonts repainting in a different typeface.
Sans serifs are overused because they look better on low resolution screens. If you haven't had to use a low resolution screen in a while then you are one of the privileged few.
At small sizes, like 8–13px, sans-serifs look better than serifs. And user interfaces and websites used to be that size.
But for 16px and up, which is what websites of today use, serifs are perfectly fine even on 1× displays (though sure, they’ll look better still on higher-DPI displays, but so will sans-serifs).
I should clarify that it depends on the font. Sans-serifs tend to have fairly even stroke width, but serifs tend to have more variable stroke width, and if the thin is too thin, you get a terrible font. That’s a common shortcoming of serif fonts, and Garamond demonstrates that, being quite unsuitable for screen use below 20px, maybe even 24px. Others like Georgia don’t suffer that weakness.
Are you sure that "websites of today" use 16px or larger universally? HN appears to use 9pt Verdana, which I believe is equivalent to 12px on my Windows system if my math is correct.
Georgia is a terribly underrated font. I'm sure it's heavily hinted to look good at small sizes, but even at large sizes it has an elegance that is lacking in e.g. Times Roman.
This sounds like it ought to be a job for pixel-density media queries [1] in CSS. I doubt this happens often if ever though, because designers. Anyone seen this approach in the wild?
I guess I specifically mean the default browser serif font. There are sites that consciously choose a serif typeface to convey some sense of "we are serious content," but you may notice that none of them use the default browser serif font (except as a last-step fallback).
If I see default browser serif, my immediate thought is "either this is an amateur or something is broken."
I don't have data to back it up, but I've understood from typography gurus that sans-serif fonts are great for signs, short blurbs, etc. But for long (multi-paragraph) reads serif fonts reduce eye fatigue. I think blogs typically fit in that category.
Georgia is a decent serif that will typically be available.
Or just `font-family: serif` which will most commonly be Times New Roman or similar, but will be whatever we prefer for those few of us that set the default fonts.
This looks exactly like what I've been searching for all this while. Do you plan to release a self hosted/open source version of this? The one thing that makes me uncomfortable about a blogging service - what happens if they shut down?
After being inspired by Bear Blog I build something similar with this exact concern in mind. Mataroa.blog [1] is a minimal blogging platform with export (to static site generators) as a first class citizen.
> You can't self host it, otherwise you'd be better off writing your own basic HTML pages or using a static site generator. You can check out the source code on github though https://github.com/HermanMartinus/bearblog
I've had a similar idea after looking for something like this to use as a log for interesting stuff that I found on the internet. I ended up building a small ruby app for this (https://github.com/adnanh/mikro) and it powers my (b)log which is still alive at http://blog.hajdarevic.net/
Since then, I've wanted to enhance it with URL unfurling, link preserving in terms of creating a snapshot of the linked URLs, smart bookmarklet etc... but never got around to doing it, as life tends to get busy, and other things take priority over that :-(
Heads up: it doesn’t have no JavaScript at present, because Cloudflare’s email protection system is in place. (I hate that thing. It mangles non-email addresses all the time if they look even vaguely like an email address, e.g. a package with version number “foo@1.2.3”. Penalising users that disable JavaScript.)
https://herman.bearblog.dev/ contains a “Get in touch” mailto: link that gets ruined, and the following script is added:
As a site owner, forcing non-JS users to click through a CAPTCHA is worth it since I get zero spam as a result. It might not be appropriate here, but it's appropriate for most personal websites.
My email address, me@chrismorgan.info, has been widely promulgated for many years. I get about five spam messages a day, and occasionally 10–15 a day for a week or two. I average around one false negative per month, and have had two false positives in the last three years (and neither mattered—in fact I kinda agreed with the spam filter’s judgement in both cases).
My email is with Fastmail, so I know that no emails are mysteriously disappearing (which seems to happen with providers like Gmail)—either the server rejects the email (so the sender will be notified), or it ends up in my mailboxes somewhere.
So my own experience is that sharing your email address publicly is harmless and useful.
Wondering how people feel about analytics on small sites such as these? Including a Google Analytics tag seems to go against the ethos of such a site, but a lot of people like to know who is visiting their blog and so including some sort of analytics should be an option.
Self hosting seems to be a decent tradeoff. Analytics services like Fathom [0], which provides a free, self-hosted solution, let you track basic statistics about your site without the data being siphoned off to some company. I have seen discussion around whether the use of analytics should even be necessary, but I feel like that's a separate conversation :)
Really cool! I like the idea to use a note taking/writing tool as a headless CMS and thought of this for Emvi [1] too. We provide access to everything through our API already... You can see usage of that here [2] and on GitHub [3].
The loading times of on Bear Blog seem a bit too high? Where does Medium for example take 16s to load?
I tried to get a better domain. I tried so hard. tiny.blog, bare.blog, bear.blog, petite.blog, smol.site, etc, etc.... Small domains are competitive if you're on a shoestring budget :P
I've not worked out when firefox reader view is visible and when it isn't, but for this blog it isn't, which is a shame because I find the chosen font unreadable.
In a nutshell, the paragraphs (<p> tags) do not contain enough text to contribute to the readability score. Reader view heuristics rewards paragraphs of at least 140 characters.
Hosting fees are pretty negligible. If it becomes a burden, then I'll cross that bridge when I get there. Thing is, if enough people are using this to make me have to upgrade my hosting, that's a great problem to have.
I noticed someone asked about self-hosting: I already had a PHP site so I wanted an easy way to integrate a blog in my site. I ended up creating a micro-blogging library[0], which just renders markdown as blog posts. The advantage is that there is no build step, you only write markdown files (VSCode has live-preview by default for markdown) and you have easy versioning as markdown is just text.
[0]: https://github.com/Cristy94/markdown-blog
I am glad that people are creating these services, but I always see a problem with this: what happens if the developer can no longer maintain the service or does not scale economically. All the blogs will end up being lost. Unless the developer is a philanthropist who plans to keep this up to the end of the day, I see these services as ideal for very specific things, not as your new home.
You can visit https://upmath.me/ and on the preview side, choose `md`. You can get latex equation as SVG URL hosted on s2cms.ru CDN. It's pretty fast and supports any website that supports markdown+images without the need for MathJax or Katex. Example post here https://katr.bearblog.dev/latex/
As mentioned in other replies--some small styling tweaks for readability would be useful and would literally only cost bytes (on the order of the x-clacks-overhead on the page...)
If you're focusing on an extremely small page size, I'd prefer an external static, cached stylesheet than inlining all styling on every page load. That and a less complicated email obfuscator. :-)
This looks minimal, but a random page that I've visited [1] duplicates article's text 6 times across different meta tags. Sure something that can be improved.
You can't self host it, otherwise you'd be better off writing your own basic HTML pages or using a static site generator. You can check out the source code on github though https://github.com/HermanMartinus/bearblog
Love to see Django apps out in the wild! It might be worth adding a COPYRIGHT file to the repo, to clarify if we do or do not have the right to run this software on our own.
Nice to see these minimal sites. I also tried to make my own blog as minimal in used bandwidth as possible without sacrificing my features. A full page load with all resources is around 22 kB.
Just a note, there seems to be an errant </div> at the end of the generated HTML.
There is definitely something to be said with respect to how bloaty many blogging packages are these days...
Actually, come to think of it, that would be a great name for a new blogging service... call it "bloaty" or "bloatly" or something like that...<g>.
It could also refer to the fact that with most blogs, the signal to noise ratio of the words comprising the articles is similarly bloated, as well... so "bloatly" would be a great name! That'll be my next Y Combinator submission for summer 2020!
"Bloatly"!
I can picture my future company's Y Combinator interview:
YC Staff: "So what does your piece of software do?"
Us: "It's a blogging platform, the very first HAAS (Hot-Air-As-a-Service) platform out there... <g> Great for politicians and salespeople! 1.21 Gigabytes of useless JavaScript per page... the Internet will end before you successfully download any page!"
YC: "So what do you call it?"
Us: "We call it Bloatly -- now give us our $150,000 check!" <g>
Unless network speeds are reversing to the modem age, browsers switching to Trident/Gecko and PCs to single core CPUs exclusively, this is 100% hipsterism.
As if using scripts or CSS is solely a choice of the CMS...
You probably don't know how many people still use slow, expensive and limited (often every part extreme) connections and old CPUs. And even on a Core i7 with a 100 Mbit/s channel many websites still feel slow.
Perhaps in rural areas in Africa and Asia. But these areas are plagued by poverty either way. Reading John Doe's blog on a 5KB web page is probably their last concern... :)
There are plenty of people in the developed US who have unreliable and slow internet. Consumers benefit from lightweight products. This is not really difficult to understand.
The comparisons are all to large content management systems. However this seems like a better fit for static site generators. Would it be better to compare to Jekyll, Hugo, etc?
Sure, serving static files will always be the fastest, but this is for people looking for a service like Medium, or Blogger, where there's no need to handle anything locally.
to be clear, you don't _need_ to do anything locally. It's possible to use static blogs without touching your site and there are lots of competing tools out there that will give you a web front end to your static blog.
For example. I use Hugo, with Forestry.io and don't have to touch my filesystem (or even my personal computer) to blog. I just go to a web page like all the CMS blogs.
I'm self-hosting rather than using something like netifly so the setup was a bit more geeky BUT it's totally doable and doesn't have to be tooo geeky.
Good idea but not sure if I would call it a blog hosting platform. My understanding is that it hosts a single parsed markdown file on a subdomain of my choosing.
Why have the same name that is linked with a well known app. 1. it confuses the user 2. No unique branding as this question is going to crop up again and again. Why the author wants to compete with this name ? Just curious
I don't think you can trademark a generic work like "Bear" .. so long as the author isn't using the logos from that app or conflating its brand, it should be fine (IANAL)
You absolutely can trademark a generic word for specific circumstances. That’s why I can’t set up a computer shop called Apple. It was also the subject of many lawsuits between Apple Computers and Apple Records.
I like it. I actually built my own platform last year in a similar vein (https://senryu.pub) but made a few different trade-offs; I allowed myself a couple of fonts (which make up most of the page size), but I can very much appreciate going even further towards minimalism. Mine's a slow-burn project that I go back to every now and then and just add small daft features just for myself, as nobody else uses the bloody thing (probably because I put a paywall on it to stop spam). Poem formatting? Absolutely.
I'm all for more platforms like this existing, I'd love a return to a legible, lightweight internet.
Also if you really want to be precise you should consider whether you're using binary prefixes vs SI prefixes, e.g. kB (10^3 bytes) vs kiB (2^10 bytes). That doesn't matter as much because the error is small for these lower values, but the casing errors definitely do matter. "mb" means millibars to me, not Megabytes!