RFC 1925(11) —
(11) Every old idea will be proposed again with a different name and a different presentation, regardless of whether it works.
Absolutely. The crawler is doing some rudimentary processing before it ever does any sort of data storage saving. That’s the sort of thing that’s being persisted behind the scenes, and it’s almost certainly both not enough to reconstruct the web page, nor is it (realistically) human-friendly. I was going to say “readable” but it’s probably some bullshit JSON or XML document full of nonsense no one wants to read.
https://fonts.google.com/specimen/Source+Code+Pro
Source code pro for quite some time.
Unrelated, but use placeholders instead of interpolation right into the query.
See: Little Bobby Tables. https://xkcd.com/327/
Wow, I would pass on this job so fast.
Not because it’s hard to fire up curl
or something, but because any company that thinks this is a better solution than a human reviewing a resume needs to be smacked. Because you know what the very next step is? They’re going to ask for a resume, and then make you sit through that bullshit where you type your resume into a hundred different boxes into their candidate management system / workday / talento / etc., and promise to “get back to you soon.”
You know how you can check if a candidate can interact with an API? Send them a coding test. Ask questions. Do some whiteboarding with them. This sort of shit is just some HR lackey ninja thinking they’re clever and edgy.
I see. It seems like you may be one of the people that try to coerce relational models into nosql stores like Dynamo.
Or course it’s possible. They even trick you into thinking it’s a good pattern by naming things “tables”.
But if you’re using Dynamo to its fullest an ORM is not going to be able to replicate that into a relational store without some fundamental changes.