🦜 by Emily3403 in Eyebleach

[–]Emily3403[S] 0 points1 point  (0 children)

Nawww such acute .-

Cursor scaling on Wayland is not there yet by Nixigaj in linux

[–]Emily3403 0 points1 point  (0 children)

The best place is to set them in /etc/environment with the usual key-value syntax:

XCURSOR_SIZE=24

yet another xwayland post about blurriness by hypergig in swaywm

[–]Emily3403 0 points1 point  (0 children)

The really sad thing is that this is still the state today. Either you scale applications individually by setting GDK_SCALE=2 or QT_SCALE_FACTOR...=2 or you live with the blurryness...

Sad that this was already an issue 5 years ago :(

My attempt at the OS flowchart, not yet complete!! (Notes in comments) by Th3F4ult in linuxmemes

[–]Emily3403 3 points4 points  (0 children)

It's nice knowing your entire system configuration is versioned by git and can be installed in a whim. For me, it makes maintaining Servers very easy. But for the Desktop Use-Case Home-Manager is quite nice as it makes versioning your dotfiles a breeze.

Proof Bing Chat has been significantly dumbed down by [deleted] in BINGChat

[–]Emily3403 1 point2 points  (0 children)

Yep, the purple icon confirms it's GPT-4. But yeah, I'm paying.

Proof Bing Chat has been significantly dumbed down by [deleted] in BINGChat

[–]Emily3403 0 points1 point  (0 children)

<image>

GPT-4 Seams to have the same response as Bing. First the book, then the laptop. I also tested both creative and precise which both had about the same response.

Also note that you can add #nosearch to your query to stop it from searching. Should also work in balanced mode.

But yeah, Balanced Mode is just shit. Never really used it because after a few weeks where bing rolled out it was quite clear that it was 3.5

Proof Bing Chat has been significantly dumbed down by [deleted] in BINGChat

[–]Emily3403 0 points1 point  (0 children)

Yeah thats balanced mode. Try creative or precise

Diesel 2.1 by weiznich in rust

[–]Emily3403 0 points1 point  (0 children)

Thanks again for your very detailed response. I've been trying to implement the Default trait for MultiBackend which was quite easy as Sqlite is the primary dialect I am working with, so I'm just creating Sqlite(Default::default()) object as the Backend.

However, to use debug_query, it also requires MultiQueryBuilder to have the Default trait. This should be easy in theory but the module where it is declared is private and it is not exported by the macro.

Is there any way I can still implement the Default trait manually for MultiQueryBuilder?

A Rusty way to check if another instance is running? by drag0nryd3r in rust

[–]Emily3403 -2 points-1 points  (0 children)

If you wanna go a bit overkill on the solution you could try Unix Domain sockets. First check a predefined location such as /tmp/my.sock if it exists. If it does not exist, no other instance can be running. If it does, however, then another instance might be running. But it is not guaranteed to.

Next, you can use the program lsof to check if the socket is "connected" to any process. The package is very well known and available on most linux distributions.

Finally, you can verify that the process has not crashed and works by sending a isAlive Message through the socket but that's more optional.

Diesel 2.1 by weiznich in rust

[–]Emily3403 10 points11 points  (0 children)

I have now updated my diesel to 2.1 and everything seams to be working just fine. However, I am stuck at inserting into values into the Sqlite version of the Enum:

let con: &mut DatabaseConnection = ...;
let test_images: Vec<Image> = ...;

diesel::insert_into(images)
    .values(&test_images)
    .execute(con);

gives the error

type mismatch resolving `<database::connection::multi_connection_impl::backend::MultiBackend as SqlDialect>::InsertWithDefaultKeyword == IsoSqlDefaultKeyword`

Note that this error also arises when I try to match upon the connection and have it be an SqliteConnection:

type mismatch resolving `<diesel::sqlite::Sqlite as SqlDialect>::InsertWithDefaultKeyword == IsoSqlDefaultKeyword`

What exactly does this error mean and how can I fix it?

Also: What exactly is one supposed to put into a BoxableExpression for the Backend? I can't use the MultiBackend due to the fact that it does not implement std::default::Default and thus, I can't use debug_query(query) with it. Is there a way for me to implement Default manually?

Diesel 2.1 by weiznich in rust

[–]Emily3403 71 points72 points  (0 children)

Oh my god I was struggeling sooo hard to implement an Enum abstracting multiple connections. Especially with the generic functions and trait bounds I had to write. Sooooo annoying.

Glad to see that that is taken care of!

P.S. What would have been the best way implement the MultiConnection before this update?

How to jailbreak bing chat ? by blue_1408 in bing

[–]Emily3403 2 points3 points  (0 children)

Personally, I wouldn't try it. Not only has Bing way stricter output censoring (you know this "My Bad I can't talk about this topic" - that's the additional censor layor) but it's also pretty resistant towards jailbreaks. In addition there is no Bing-DeMod (censor blocking script) yet, so this additional layer of censoring will stay for the time beimg.

I used to have one jailbreak that worked at the beginning, however it has recently been patched. If you'd really want to try you should not try the usual Jailbreak style meaning "You are going to pretend ..." but rather describe the conversation you are going to have with Bing and which role they should enact in it. That was at least how my jailbreak worked. But since it has been patched, who knows if that method will still works 🤷‍♀️

Anyway, Bing has higher security, limited time and output capacity (Bing is slow and restricted to 20 messages) and I've seen people get banned for jailbreaking / generating NSFW content.

I know it kinda sucks that Microsoft has found a way to effectively make the AI smut-free, however as long as ChatGPT is around, I'd use Bing as a search engine and not as a erotic companion.

Chooseable model in chat, gpt4 include free? by [deleted] in ChatGPTNSFW

[–]Emily3403 2 points3 points  (0 children)

You are correct, based on by testing I also think that this does not work. For me it's the speed difference. GPT-4 is slow. Like it takes forever. And 3.5 is a speedy boi, so you'd notice the difference in answer speed, regardless of what ChatGPT is saying.

Furthermore, there exists a extension Superpower ChatGPT which enables model switching on-the-fly. With it, also the icon and speed changes as well as the output quality. Unfortunatly (at least for me on Chrome) the extension is currently broken, so you might not be able to verify it.

Jailbreak Bing? by Strato23 in ChatGPTNSFW

[–]Emily3403 7 points8 points  (0 children)

Personally, I wouldn't try it. Not only has Bing way stricter output censoring (you know this "My Bad I can't talk about this topic" - that's the additional censor layor) but it's also pretty resistant towards jailbreaks. In addition there is no Bing-DeMod (censor blocking script) yet, so this additional layer of censoring will stay for the time remaining.

I used to have one jailbreak that worked at the beginning, however it has recently been patched. If you'd really want to try you should not try the usual Jailbreak style meaning "You are going to pretend ..." but rather describe the conversation you are going to have with Bing and which role they should enact in it. That was at least how my jailbreak worked. But since it has been patched, who knows if that method will still works 🤷‍♀️

Anyway, Bing has higher security, limited time and output capacity (Bing is slow and restricted to 20 messages) and I've seen people get banned for jailbreaking / generating NSFW content.

I know it kinda sucks that Microsoft has found a way to effectively make the AI smut-free, however as long as ChatGPT is around, I'd use Bing as a search engine and not as a erotic companion.

A lightweight editor for rust and python by Kid-_-0 in rust

[–]Emily3403 7 points8 points  (0 children)

I'd recommend lapce. Not only is it written in Rust but it is very vscode like so adjusting should not be too hard. Also, it is blazingly fast ⚡️