TheHobbyist

joined 1 year ago
[–] TheHobbyist -1 points 1 week ago* (last edited 1 week ago) (4 children)

I think these kind of situations are where ZFS snapshots shine: you're back in a matter of seconds with no data loss (assuming you have a recent snapshot before the mistake).

Edit: yeah no, if you operate at the disk level directly, no local ZFS snapshot could save you...

[–] TheHobbyist 11 points 1 week ago

Really nice to be able to give a call link to another signal user without having to exchange phone numbers or even username. Quite empowering!

One of the features I'd love to see on Signal mobile is screen sharing. Other than that I'm quite satisfied with signal calls all together.

[–] TheHobbyist 18 points 2 weeks ago (10 children)

Indeed, totally an Apple approach to modularity: it is a proprietary Apple SSD...

[–] TheHobbyist 1 points 2 weeks ago
[–] TheHobbyist 11 points 2 weeks ago

I see, so there is indeed a broader context to the burning alone, it was also with additional verbal hatred and then possibly the location, and the overall intention. I think this makes it clearer. Thanks

[–] TheHobbyist 28 points 2 weeks ago (15 children)

Not familiar with the guy himself who maybe does deserve criticism and prison, but about the Quran burning, is it genuinely fair to sentence someone to prison for that? Is it equivalent to burning the cross? The Swedish flag? I might be mission a broader context, but I don't feel like someone burning my symbol or flag should be punished with prison. Am I alone? I would hate it, don't get me wrong, but I still feel it goes in freedom of expression.

[–] TheHobbyist 1 points 2 weeks ago* (last edited 2 weeks ago)

The proud dad's name ends with Unis and the kid remembers the X first digits, hence Unix, hence Linux!

[–] TheHobbyist 1 points 2 weeks ago

They do mention compatibility a lot, if it's hardware, I agree with you. But perhaps they mean something else?

[–] TheHobbyist 51 points 2 weeks ago

This. I will resume my recommendation of Bitwarden.

[–] TheHobbyist 3 points 2 weeks ago (1 children)

I didn't say it can't. But I'm not sure how well it is optimized for it. From my initial testing it queues queries and submits them one after another to the model, I have not seen it batch compute the queries, but maybe it's a setup thing on my side. vLLM on the other hand is designed specifically for the multi co current user use case and has multiple optimizations for it.

[–] TheHobbyist 23 points 2 weeks ago* (last edited 2 weeks ago) (5 children)

I run the Mistral-Nemo(12B) and Mistral-Small (22B) on my GPU and they are pretty code. As others have said, the GPU memory is one of the most limiting factors. 8B models are decent, 15-25B models are good and 70B+ models are excellent (solely based on my own experience). Go for q4_K models, as they will run many times faster than higher quantization with little performance degradation. They typically come in S (Small), M (Medium) and (Large) and take the largest which fits in your GPU memory. If you go below q4, you may see more severe and noticeable performance degradation.

If you need to serve only one user at the time, ollama +Webui works great. If you need multiple users at the same time, check out vLLM.

Edit: I'm simplifying it very much, but hopefully should it is simple and actionable as a starting point. I've also seen great stuff from Gemma2-27B

Edit2: added links

Edit3: a decent GPU regarding bang for buck IMO is the RTX 3060 with 12GB. It may be available on the used market for a decent price and offers a good amount of VRAM and GPU performance for the cost. I would like to propose AMD GPUs as they offer much more GPU mem for their price but they are not all as supported with ROCm and I'm not sure about the compatibility for these tools, so perhaps others can chime in.

Edit4: you can also use openwebui with vscode with the continue.dev extension such that you can have a copilot type LLM in your editor.

[–] TheHobbyist 3 points 2 weeks ago* (last edited 2 weeks ago)

As you probably know, an LLM works iteratively: you give it instructions and it "auto-completes", one token at a time. Every time you want to generate the next token, you have to perform the whole inference task, which is expensive.

However, verifying if a next token is the correct one, can be cheap because you can do it in parallel. For instance, take the sentence " The answer to your query is that the sky is blue due to some physical concept". If you wanted to check whether your model would output each one of those tokens, you would split the sentence after every token and you could batch verify the next token for every split and see whether the next token matches the sentence.

Speculative decoding is the process where a cheap and efficient draft model is used to generate a tentative output, which is then verified in parallel by the expensive model. Because the cheap draft model is many times quicker, you can get a sample output very fast and batch verify the output with the expensive model. This saves a lot of computational time because all the parallel verifications require a single forward pass. And the best part is that it has zero effect on the output quality of the expensive model. The cost is that you know have to run two models, but the smaller one may be a tenth of the size, so runs possibly 10x faster. The closer the draft model output matches the expensive model output, the higher the inference speed gain potential.

 

Hi folks, I'm looking for a specific YouTube video which I watched around 5 months ago.

The gist of the video is that it was comparing the transcoding performance of an Intel iGPU when used natively, compared to when passed through to a VM. From what I recall there was a significant performance hit and it was around 50% or so (in terms of fps transcoding). I believe the test was performed on jellyfin. I don't remember whether it was using xcpng, proxmox or another OS. I don't remember which channel published this video nor when it was published, just that I watched it sometime between April and June this year.

Anyone recall or know what video I'm talking about? Possible keywords include: quicksync, passthrough, sriov, iommu, transcoding, iGPU, encoding.

Thank you in advance!

 

Hi y'all,

I am exploring TrueNAS and configuring some ZFS datasets. As ZFS provides with some parameters to fine-tune its setup to the type of data, I was thinking it would be good to take advantage of it. So I'm here with the simple task of choosing the appropriate "record size".

Initially I thought, well this is simple, the dataset is meant to store videos, movies, tv shows for a jellyfin docker container, so in general large files and a record size of 1M sounds like a good idea (as suggested in Jim Salter's cheatsheet).

Out of curiosity, I ran Wendell's magic command from level1 tech to get a sense for the file size distribution:

find . -type f -print0 | xargs -0 ls -l | awk '{ n=int(log($5)/log(2)); if (n<10) { n=10; } size[n]++ } END { for (i in size) printf("%d %d\n", 2^i, size[i]) }' | sort -n | awk 'function human(x) { x[1]/=1024; if (x[1]>=1024) { x[2]++; human(x) } } { a[1]=$1; a[2]=0; human(a); printf("%3d%s: %6d\n", a[1],substr("kMGTEPYZ",a[2]+1,1),$2) }'

Turns out, that's when I discovered it was not as simple. The directory is obviously filled with videos, but also tiny small files, for subtitiles, NFOs, and small illustration images, valuable for Jellyfin's media organization.

That's where I'm at. The way I see it, there are several options:

    1. Let's not overcomplicate it, just run with the default 64K ZFS dataset recordsize and roll with it. It won't be such a big deal.
    1. Let's try to be clever about it, make 2 datasets, one with a recordsize of 4K for the small files and one with a recordsize of 1M for the videos, then select one as the "main" dataset and use symbolic links for each file to the other dataset such that all content is "visible" from within one file structure. I haven't dug too much in how I would automate it, but might not play nicely with the *arr suite? Perhaps overly complicated...
    1. Make all video files MKV files, embed the subtitles, rename the videos to make NFOs as unnecessary as possible for movies and tv shows (though this will still be useful for private videos, or YT downloads etc)
    1. Other?

So what do you think? And also, how have your personally set it up? Would love to get some feedback, especially if you are also using ZFS and have a videos library with a dedicated dataset. Thanks!

Edit: Alright, so I found the following post by Jim Salter which goes through more detail regarding record size. It clarifies my misconception about recordsize not being the same as the block size, but also it can easily be changed at any time. It's just the size of the chunks of data to be read. So I'll be sticking to 1M recordsize and leave it at that despite having multiple smaller files, because the important will be to effectively stream the larger files. Thank you all!

 

Sean who reviewed the Laptop 16 from theVerge and was unable to properly test the Graphics Module due to the ventilation being broken, was invited to Framework HQ to test a fully functional version.

There is hardly any new information other than that it worked, but was running at maximum fan speed due to fan curves not having been implemented yet.

He was able to play Cyberpunk 2077 and Halo Infinite on high for 15 min without throttling.

Nothing more than what was expected but it's good to get a confirmation.

 

Dave2d who's been supportive of Framework preordered the Laptop 16.

He's a bit concerned about the pricing and questions the upgradability of the Laptop 16 specifically.

Personally I understand his point, but I think the upgradability alone is probably not a good reason to buy the Laptop 16. It's always been a package, which includes:

  • repairability
  • modularity
  • support of the movement/mission
  • the versatility of reusing parts for other use cases (e.g. the motherboard as thin-client)
  • a laptop that actually does not have Linux as an afterthought
  • the openness with the expansion card and (hopefully expansion bay) ecosystem
  • and maybe even more?

It's true that the laptop is expensive when you compare specs for specs but that was not the reason to buy it either. Do I wish it was cheaper? You bet. But like with all new startups, if it works out, if it scales, prices could come down. Long live Framework!

 

The verge got a hands on with the Framework Laptop 16 and wrote an article and published a YouTube video.

Article here: https://www.theverge.com/22665800/framework-laptop-16-hands-on-preview-modular-gaming-laptop

Video here: https://www.youtube.com/watch?v=7xq8rOlwW5Y

Piped link: https://www.piped.kavin.rocks/watch?v=7xq8rOlwW5Y

 

This is a great surprise, the pre-orders are open before the end of the Laptop 16 deep dives.

Quoting the blog post below:

We’re excited to share that Framework Laptop 16 pre-orders are now open, with configurations powered by the latest AMD Ryzen™ CPUs and AMD Radeon™ GPUs. This is truly a notebook like no other: thin and refined, while empowering you with desktop PC-level customization, repairability, and upgradability, including a fully reconfigurable input deck and modular discrete graphics. Prices start at $1399 USD for DIY Edition and $1699 USD for pre-built systems with Ryzen™ 7 7840HS, and adding an AMD Radeon™ RX 7700S Graphics Module brings starting prices to $1799 and $2099 USD.

Pre-orders that include a Graphics Module with an eligible AMD Radeon™ GPU will receive a free download code for one of the biggest games of the year: Starfield™ Premium Edition. Quantities are limited*, and we’ll be sending out the code prior to the game’s early access launch.

As always, we’re following a batch ordering system, with the first batches shipping in Q4 2023. Ordering is open now in all of our current countries: US, Canada, Germany, France, UK, Netherlands, Austria, Ireland, and Australia. A fully refundable $100 deposit is all you need to get in line. We recommend getting your order in early if you’d like to get a system this year. We’re sharing much more detail today to help you decide if this is your next (and maybe final?) laptop.

We’re not only using AMD Ryzen™ and Radeon™ silicon, but we developed this product in close collaboration with AMD as part of the AMD Advantage program. We’re leveraging AMD Ryzen™ 7040 Series processors, the latest generation that we also use in Framework Laptop 13 (AMD Ryzen™ 7040 Series), this time with HS-class parts optimized for gaming and creation. Configurations start with the Ryzen™ 7 7840HS with 8 Zen 4 CPU cores at up to 5.1GHz boost, and we also offer the totally overkill, top of the line Ryzen™ 9 7940HS with up to 5.2GHz boost. We worked with Cooler Master to design a thermal system with dual 75mm fans, three heatpipes, and a liquid metal thermal interface, enabling 45W continuous processor load while also keeping the laptop cool and quiet. There’s fantastic graphics performance built in too, with Radeon 780M graphics with 12 RDNA 3 cores, capable of running a range of modern game titles.

If you want substantially more graphics horsepower, Framework Laptop 16 delivers the holy grail for high performance notebooks: optional discrete graphics using our new Expansion Bay system, allowing generation-over-generation graphics upgradeability. The first Graphics Module for the Expansion Bay features the AMD Radeon™ RX 7700S GPU. We’ve maxed out the capabilities of the chip, with 100W sustained TGP and 8GB GDDR6 at up to 18Gbps. Because the Graphics Module contains its own dedicated heatsink and higher CFM fans, both the CPU and GPU can run at full wattage simultaneously when needed. This GPU excels for both work and play, with 32 compute units at up to 2.2GHz, enabling high-end gaming, incredible rendering and encoding throughput, and excellent acceleration for AI and other applications.

Of course, it’s not enough to have great silicon. A high-performance laptop demands thoughtful integration across every subsystem. We leverage a custom 16 inch 2560x1600 display, supporting 165Hz with FreeSync, 500 nit, 1500:1 contrast, and 100% DCI-P3 color gamut, making it excel across gaming, creation, and productivity. The 85Wh battery lasts you through a full workday, retains typically 80% capacity after 1,000 cycles, and is easy to replace if ever needed. Quad speakers connected to a smart amp provide high fidelity audio across a wide frequency band. For connectivity, we enabled WiFi 6E and Bluetooth 5.2 using AMD’s new RZ616 M.2 module. We built in a 1080p webcam with dual mics and hardware privacy switches, and for security we incorporated a Windows and Linux-compatible fingerprint reader.

For I/O, we brought in the Expansion Card system that enables full customization of port selection, with three slots on each side. The rear two support USB4, the middle left handles USB 3.2+DisplayPort output, and the remaining three have USB 3.2. The back two slots on each side can take up to 240W power input over USB-C using USB-PD 3.1. We offer a compact, ultra high efficiency 180W GaN adapter with detachable cables, and with DIY Edition, you can choose to bring your own.

We advanced laptop industrial design on both form and function, combining a refined form factor and unprecedented levels of customization. The Framework Laptop 16 is 17.95mm thick and 2.1kg (4.6lbs), going to 20.95mm in the back section and 2.4kg (5.3lbs) with a Graphics Module inserted. The chassis is made of robust and lightweight thixomolded magnesium alloy and CNC aluminum enclosure parts. As always for Framework products, user-friendly design goes below the surface too, with every internal module simple to replace or upgrade, including the Mainboard for generational processor upgrades.

The input system is fully hot-swappable using Input Modules, letting you reconfigure between a centered keyboard or offset with a numpad. The keyboard and numpad look and feel excellent, with 1.5mm key travel, optional per-key RGB, NKRO, and fully open source QMK firmware. Input deck personalization goes even further, with Spacers in a range of colors, a programmable LED Matrix module, and an RGB Macropad all available as options. We’ve open sourced this system to enable third party and community development too, and we can’t wait to see the insanely cool modules that come from that.

When ordering a Framework Laptop 16, you can choose between pre-built options that are ready to go out of the box with Windows 11 or the DIY Edition that you can configure more deeply, assemble yourself, and bring your preferred OS, including Linux. AMD has a strong focus on Linux drivers, and we provide in-house support and guides for Ubuntu LTS and Fedora. At order time for both pre-built and DIY Edition, you can choose your Input Modules, Expansion Cards, and Expansion Bay Modules. DIY Edition additionally lets you pick your Bezel color, memory (up to 64GB of DDR5-5600), storage (two M.2 NVMe drives), and power adapter. As always, you’ll be able to pick up additional modules or upgrades in the Framework Marketplace whenever you need.

With the Framework Laptop 16, we’re taking our mission to the next level with a sleek, portable system that has the flexibility and generational upgradeability of a full desktop rig. This redefines what a high performance laptop can be: a machine that is uniquely yours to mold to your needs and use for as long as you’d like. We can’t wait to see what you do with it.

*The free Starfield™ Premium Edition download code is a limited time and quantity offer. You may lose eligibility for this pre-order gift if you make certain order modifications, such as removing the Graphics Module from your pre-order. All canceled pre-orders will no longer be eligible to receive the free game code. For terms and conditions, see www.amdrewards.com/terms.

 

I'm curious and am playing around with a new EDA tool and am looking at practicing by designing a PCB which should be roughly 28x26mm footprint (give or take a few mm...).

It should be an LTE cat 4 device, connected by USB type C for the framework laptop and is unlikely to include antennas.

Where I struggle is identifying potential modems to use. The only one even remotely close is the u-blox LARA-L6, which is 24x26mm. What alternatives are there?

I am trying to see what gets sold in these USB dongles but there is little info. The few I have identified seem to make use of the Qualcomm 9207, but its's unclear to me if its a ready chip (which is what the MDM9207 is?) Or if it is an IP core to integrate in one's own chip?

A video I came across seem to indicate it (the MDM version) is tiny:

https://www.youtube.com/watch?v=ToCyUCIoXEM at 2:13

But will probably needadditional things to be integrated and I created an account at Qualcomm but they won't give anything unless I'm certified from a company to be a customer and actually integrate it...

view more: ‹ prev next ›