• 16 Posts
  • 98 Comments
Joined duela 2 urte
cake
Cake day: abu. 24, 2020

help-circle
rss

Most of websites are built on Wordpress, so thats a start.

There is a lot of JS around, but most websites have some Open Source foundation.



The same way you can get obsessed with… anything really.

At the time i was obssesed because many animes were so interesting and creative, and exciting.


Yo utilizo criptos con muy poco valor de comisión como Monero para usarla como Monedas.

La gente me sigue diciendo que invierta y haga trading, pero no me interesa, prefiero usarlas como verdaderas monedas, para pagar servicios en línea que de otra forma me serían imposibles, debido a que no tengo una tarjeta de crédito internacional, y esas son muy difíciles de conseguir en mi país.


Our reality is the real meme


Why can’t we have both?


And yet aren’t able to shut down all the CP on this fucking platform, just yesterday an idiot came into a group I manage openly asking for CP. If telegram is gonna use mass surveillance, at least make it useful god damn.


The US will push you to be homeless, and then jail you for that.


What the hell did i just read?


We know that has been happening for so long. Tho is kinda scary.



I know nothing about lol (except for the music wich is always dope) I tried to play many years ago but didnt go nowhere

Yet I absolutely love arcane. The animation, the story, everything is so fucking perfect. And I love Jinx, 100% transition goal.

I wanna know more about the lore, it looks so good



At this point they should just ask help to Anthony since Linus has no idea what he is doing. Although Luke is having a better experience and not too many complains. In a recent WAN show Luke said that he likes Linux mire for working since “it gets less in the way that windows”



  • piracy.

I wont pay nothing. Popcorntime already gives you a Netflix like interface, or just put ur collection on Kodi and it looks great.

I watch Netflix from my sister’s account lol, I pirate everything else


Wait, then what about fucking ass? Will the dick disappear?



Im glad they are working on it. It looks great.

At least Conversations on android kinda looks like WA and its easy to use



Smoke what? Me who just ran out of vaper essence


Is there some app that allows me to convert markdown to PDF on the go?
Hi. I use Joplin and quillnote to make notes on the go. But many times I need then as PDF. On the computer there in no problem but none of these allow me to export as PDF on my phone. Is there some app that allows me to do this?
fedilink


What opinions you have on the Facebook Rebranding?
You think is Facebook trying to just wash their hands as Facebook itself will become part of a bigger conglomerate created by themselves?
fedilink


Scribe.rip Open source Medium Front end
This is a recent toot from the creator: I've made an alternative frontend to Medium: https://scribe.rip If you, like me, are occasionally forced to read articles on medium.com, now you can read them on Scribe instead! Feedback welcome. @edwardloveall@mastodon.technology
fedilink

Is Silence still maintained?
I wanted to install Silence, but fdroid says it was last updated 2 year ago. Is the project still alive and active? I remember using it a while ago and i really liked it. That's why i want to install it again.
fedilink

What launcher you recommend the most on fdroid?
Hi. Im very happy right now with smart launcher. But I would like something fully open source. I want something minimal, but that allows me to have desktop pictures, because I make my own wallpapers. Also it would be good to create folders or apo categories. Thank you.
fedilink

Do you think the fediverse has become a save space for radicals?
The alt social media are a good place to run away from algorithms and explicit mass manipulation from other big tech, be more free, independent, and spread ideas, but I also think that the fediverse is being filling with people with radical political ideologies, some kicked out of traditional social platforms for this same reason. I'm not saying that we can't have discussions, but I think that many people are making their own echo Chambers in the fediverse.
fedilink


“Trump Accuses Google of Burying Conservative News in Search Results,” reads an August 28 New York Times headline. The piece features a bombastic president, a string of bitter tweets, and accusations of censorship. “Algorithms” are mentioned, but not until the twelfth paragraph. Trump—like so many other politicians and pundits—has found search and social media companies to be convenient targets in the debate over free speech and censorship online. “They have it RIGGED, for me & others, so that almost all stories & news is BAD,” the president recently tweeted. He added: “They are controlling what we can & cannot see. This is a very serious situation---will be addressed!” Trump is partly right: They are controlling what we can and cannot see. But “they” aren’t the executives leading Google, Facebook, and other technology companies. “They” are the opaque, influential algorithms that determine what content billions of internet users read, watch, and share next. These algorithms are invisible, but they have an outsized impact on shaping individuals’ experience online and society at large. Indeed, YouTube’s video-recommendation algorithm inspires 700,000,000 hours of watch time per day—and can spread misinformation, disrupt elections, and incite violence. Algorithms like this need fixing. But in this moment, the conversation we should be having—how can we fix the algorithms?—is instead being co-opted and twisted by politicians and pundits howling about censorship and miscasting content moderation as the demise of free speech online. It would be good to remind them that free speech does not mean free reach. There is no right to algorithmic amplification. In fact, that’s the very problem that needs fixing. TO SEE HOW this algorithm amplification works, simply look to RT, or Russia Today, a Russian state-owned propaganda outlet that’s also among the most popular YouTube presences. RT has amassed more than 6 billion views across 22 channels, more than MSNBC and Fox News combined. According to YouTube chief product officer Neal Mohan, 70 percent of views on YouTube are from recommendations—so the site’s algorithms are largely responsible for amplifying RT’s propaganda hundreds of millions of times. How? Most RT viewers don’t set out in search of Russian propaganda. The videos that rack up the views are RT’s clickbait-y, gateway content: videos of towering tsunamis, meteors striking buildings, shark attacks, amusement park accidents, some that are years old but have comments from within an hour ago. This disaster porn is highly engaging; the videos have been viewed tens of millions of times and are likely watched until the end. As a result, YouTube’s algorithm likely believes other RT content is worth suggesting to the viewers of that content—and so, quickly, an American YouTube user looking for news finds themselves watching Russia’s take on Hillary Clinton, immigration, and current events. These videos are served up in autoplay playlists alongside content from legitimate news organizations, giving RT itself increased legitimacy by association. The social internet is mediated by algorithms: recommendation engines, search, trending, autocomplete, and other mechanisms that predict what we want to see next. The algorithms don’t understand what is propaganda and what isn’t, or what is “fake news” and what is fact-checked. Their job is to surface relevant content (relevant to the user, of course), and they do it exceedingly well. So well, in fact, that the engineers who built these algorithms are sometimes baffled: “Even the creators don’t always understand why it recommends one video instead of another,” says Guillaume Chaslot, an ex-YouTube engineer who worked on the site’s algorithm. These opaque algorithms with their singular purpose—“keep watching”—coupled with billions of users is a dangerous recipe. In recent years, we’ve seen how dire the consequences can be. Propaganda like RT content is circulated far and wide to disinform and worsen polarization, especially during democratic elections. YouTube’s algorithms can also radicalize by suggesting “white supremacist rants, Holocaust denials, and other disturbing content,” Zeynep Tufekci recently wrote in the Times. “YouTube may be one of the most powerful radicalizing instruments of the 21st century.” The problem extends beyond YouTube, though. On Google search, dangerous anti-vaccine misinformation can commandeer the top results. And on Facebook, hate speech can thrive and fuel genocide. A United Nations report about the genocide in Myanmar reads: “The role of social media is significant. Facebook has been a useful instrument for those seeking to spread hate, in a context where for most users Facebook is the Internet … The extent to which Facebook posts and messages have led to real-world discrimination and violence must be independently and thoroughly examined.” So what can we do about it? The solution isn’t to outlaw algorithmic ranking or make noise about legislating what results Google can return. Algorithms are an invaluable tool for making sense of the immense universe of information online. There’s an overwhelming amount of content available to fill any given person’s feed or search query; sorting and ranking is a necessity, and there has never been evidence indicating that the results display systemic partisan bias. That said, unconscious bias is a concern in any algorithm; this is why tech companies have investigated conservative claims of bias since the Facebook Trending News debacle of 2016. There hasn’t been any credible evidence. But there is a trust problem, and a lack of understanding of how rankings and feeds work, and that allows bad-faith politicking to gain traction. The best solution to that is to increase transparency and internet literacy, enabling users to have a better understanding of why they see what they see—and to build these powerful curatorial systems with a sense of responsibility for what they return. There have been positive steps in this direction. The examples of harms mentioned above have sparked congressional investigations aimed at understanding how tech platforms shape our conversations and our media consumption. In an upcoming Senate hearing next week, the Senate Intelligence Committee will ask Jack Dorsey of Twitter and Sheryl Sandberg of Facebook to provide an accounting of how, specifically, they are taking steps to address computational propaganda. It’s imperative that we focus on solutions, not politics. We need to build on those initial investigations. We need more nuanced conversations and education about algorithmic curation, its strange incentives, and its occasionally unfortunate outcomes. We need to hold tech companies accountable—for irresponsible tech, not evidence-free allegations of censorship—and demand transparency into how their algorithms and moderation policies work. By focusing on the real problem here, we can begin addressing the real issues that are disrupting the internet—and democracy.
fedilink



Dark Web Users. What good use you give to it and why?
Hi. I'm getting interested again on the deep web, but I wanna know what is the best use i can give to it, outside of every single illegal content stuff, because I don't want any of that. I wanna know the good side of it.
fedilink


What about creating porn images communities?
Like... come on, we have to use the NSFW tag for something.
fedilink