I predict that there be a movement from 4k displays to 8k. Rather it will be a movement from 4k to vr based headsets. There's no point to anything higher resolution than 4k unless the objective is a 42in display from 3'. The people who can't use vr will be stuck on the panel displays and I'm sure cable companies will take advantage of augmented reality experience, I think 360° tv will be standard with vr integration as well. We'll also start to see a decrease in the scale of memory increases on GPUs as well. It'll probably top off at 32 GB within 10 years with cpu and GPU cores more reliant than memory.
What do you guys think?
[url]http://www.anandtech.com/show/10525/ten-year-anniversary-of-core-2-duo-and-conroe-moores-law-is-dead-long-live-moores-law/1[/url]
Let's take it a step at a time and say vr probably won't be just a gimmick
[QUOTE=nikomo;50801426]
[QUOTE]I told this to the AfterPay representative at the GBSB event on 27th July 2016. The
representative told me that the problem is already known. [B]The only concrete action
taken was to threaten my removal from the GBSB event[/B].[/QUOTE]
I like how people are stupid enough to still do this shit.[/QUOTE]
What the actual fuckity fuck?
"hey thanks for letting us know, [I]BTW [/I]you're in trouble now for telling us about it".
Went to the UPS store, they took down my info and said they buyer would be contacted within 9 days to inspect the packaging and item.
[QUOTE=TheDestroyerOfall;50801659]I predict that there be a movement from 4k displays to 8k. Rather it will be a movement from 4k to vr based headsets. There's no point to anything higher resolution than 4k unless the objective is a 42in display from 3'. The people who can't use vr will be stuck on the panel displays and I'm sure cable companies will take advantage of augmented reality experience, I think 360° tv will be standard with vr integration as well. We'll also start to see a decrease in the scale of memory increases on GPUs as well. It'll probably top off at 32 GB within 10 years with cpu and GPU cores more reliant than memory.
What do you guys think?
[url]http://www.anandtech.com/show/10525/ten-year-anniversary-of-core-2-duo-and-conroe-moores-law-is-dead-long-live-moores-law/1[/url][/QUOTE]
I think its inevitable we move to 8k, even TV will (The cameras already exist and have been used to film the Olympics in 2012) but probably not for a good while, 4K is bad enough as it is and now we will have to look at rendering 4X that and I don't see it happening currently. I'm personally yet to be convinced to move to even 4K would benefit my video games as much as 1080p did and would rather have a 1080p120 screen which would also give me the benefit of supporting 30FPS and 24FPS without dodgy framerate pulldown conversion.
[QUOTE=TheDestroyerOfall;50801659]I predict that there be a movement from 4k displays to 8k. Rather it will be a movement from 4k to vr based headsets. There's no point to anything higher resolution than 4k unless the objective is a 42in display from 3'. The people who can't use vr will be stuck on the panel displays and I'm sure cable companies will take advantage of augmented reality experience, I think 360° tv will be standard with vr integration as well. We'll also start to see a decrease in the scale of memory increases on GPUs as well. It'll probably top off at 32 GB within 10 years with cpu and GPU cores more reliant than memory.
What do you guys think?
[url]http://www.anandtech.com/show/10525/ten-year-anniversary-of-core-2-duo-and-conroe-moores-law-is-dead-long-live-moores-law/1[/url][/QUOTE]
As far as games and graphics go there is still plenty of resolution improvements even after 8k before you actually start getting no returns. Sure it has diminishing effects, but so does increasing the memory or computing power.
[img]http://cld.moe/files/2016-07-30_22-28-26.png[/img]
I think I figured out why my macbook has been reduced to a crawl lately
[img]http://i.imgur.com/PjzKTQU.png[/img]
[editline]30th July 2016[/editline]
guys i force closed it and not it doesn't start again wtf
[QUOTE=Confuzzed Otto;50802248]I think I figured out why my macbook has been reduced to a crawl lately
[img]http://i.imgur.com/PjzKTQU.png[/img]
[editline]30th July 2016[/editline]
guys i force closed it and not it doesn't start again wtf[/QUOTE]
There's a reason many just use the spotify web-player nowadays.
Got my first bug submission to Arq. The little limiter works but looks like it's per thread, not total: [img]http://i.imgur.com/bjRg6rH.png[/img]
Basically it was running at like an inverse bit-to-byte conversion. I had it set to 135KB/s and it was uploading at like 1MB/s, not 1Mbps.
[QUOTE=Van-man;50802274]There's a reason many just use the spotify web-player nowadays.[/QUOTE]
Does it stream in the highest quality?
[QUOTE=Confuzzed Otto;50802303]Does it stream in the highest quality?[/QUOTE]
No idea since I don't personally use Spotify after they ditched sorting songs by user rating (without using a convoluted hack), but it's pretty clear that their clients have turned to shit even more.
[QUOTE=Genericenemy;50802228]I think its inevitable we move to 8k, even TV will (The cameras already exist and have been used to film the Olympics in 2012) but probably not for a good while, 4K is bad enough as it is and now we will have to look at rendering 4X that and I don't see it happening currently. I'm personally yet to be convinced to move to even 4K would benefit my video games as much as 1080p did and would rather have a 1080p120 screen which would also give me the benefit of supporting 30FPS and 24FPS without dodgy framerate pulldown conversion.[/QUOTE]
Back in the good old days it was always the 2D stuff that got the stupid high resolution stuff first. 3D came along later, initially in the really high-end but eventually being found in commodity hardware.
[QUOTE=Cold;50802243]As far as games and graphics go there is still plenty of resolution improvements even after 8k before you actually start getting no returns. Sure it has diminishing effects, but so does increasing the memory or computing power.
[img]http://cld.moe/files/2016-07-30_22-28-26.png[/img][/QUOTE]
I never really liked this image. It looks like they just applied a smooth modifier to the third one and didn't really add any more detail. The clothes look too rounded so those extra 10x triangles aren't really useful at all.
It just seems to make it look like it's the poly counts fault to attempt to drive home the idea of diminishing returns.
[QUOTE=Snickerdoodle;50802444]I never really liked this image. It looks like they just applied a smooth modifier to the third one and didn't really add any more detail. The clothes look too rounded so those extra 10x triangles aren't really useful at all.
It just seems to make it look like it's the poly counts fault to attempt to drive home the idea of diminishing returns.[/QUOTE]
[url=https://youtu.be/Z45nbzMLk98?t=63]I can fix that.[/url] :v:
[QUOTE=Confuzzed Otto;50802248]I think I figured out why my macbook has been reduced to a crawl lately
[img]http://i.imgur.com/PjzKTQU.png[/img]
[editline]30th July 2016[/editline]
guys i force closed it and not it doesn't start again wtf[/QUOTE]
It says there's updates pretty often though, so not sure how you even managed to leak for that long (even the leaking part is confusing for me as well, I use it on Mac & Windows and they work perfectly fine with no leaks on extreme)
[QUOTE=AncientFryup;50800570]The Spectre 13 I think it's known as, it's just the standard ultrabook not the x360.[/QUOTE]
I was using the HP Spectre x360 at one point, it was a decent ultrabook and the track pad was really wide. I actually liked it because it made scrolling relatively easy when you're sitting down with it on your lap, or laying on your bed with it on your chest/belly. Battery lasted relatively long too so I'd imagine the Spectre 13 would be good too.
Fucking MySQL needs to stop eating my RAM.
[QUOTE=rhx123;50801525]I would avoid it out of principal because of HPs unnecessary charger DRM on the Type C ports. The XPS 13 with Iris Graphics is your best bet.[/QUOTE]
Charger DRM.. on the fucking Type C ports.. really HP..
[QUOTE=garychencool;50802922]Charger DRM.. on the fucking Type C ports.. really HP..[/QUOTE]
HP have a burning hatred for third party chargers.
I had to gut a (dead) official charger for it's ID chip so it'll work with my universal car charger for my old Pavillion DV7.
The bonus was that I could also use the official chargers cable which was also of better quality than the universal one's adapter.
[URL="https://facepunch.com/showthread.php?t=1528886"]People liked the review!!![/URL]
This makes me happy. I've been installing an entire small office building's cat5 network in the hot and sweaty ceiling all day so coming home to this is pretty awesome.
[QUOTE=Snickerdoodle;50803267][URL="https://facepunch.com/showthread.php?t=1528886"]People liked the review!!![/URL]
This makes me happy. I've been installing an entire small office building's cat5 network in the hot and sweaty ceiling all day so coming home to this is pretty awesome.[/QUOTE]
Game reviews can be therapeutic, even more so when people feel the same way.
My FO4 video took a while to pick up steam:
[media]https://www.youtube.com/watch?v=So8Br2fmSYA[/media] But seems like the more people played the game, the more people agreed with my points. The longer people play FO4, the more they realized it's terrible.
I'm not sure if my review will pick up a whole lot of Steam for that reason. I Love Katamari is actually abandonware at this point :x
Fuck me, I dropped my Oneplus One and the glass is shattered. Time to buy a new digitizer (I guess with an LCD) and try to replace it. Seems to be pretty hard to replace it though, have to do my best to not smash everything.
[QUOTE=tratzzz;50803321]Fuck me, I dropped my Oneplus One and the glass is shattered. Time to buy a new digitizer (I guess with an LCD) and try to replace it. Seems to be pretty hard to replace it though, have to do my best to not smash everything.[/QUOTE]
Good luck, phone repair was stressful but not as difficult as I thought it would be
[QUOTE=Confuzzed Otto;50802248]I think I figured out why my macbook has been reduced to a crawl lately
[img]http://i.imgur.com/PjzKTQU.png[/img]
[editline]30th July 2016[/editline]
guys i force closed it and not it doesn't start again wtf[/QUOTE]
i actually had to make a cron job that would restart onedrive daily because it keeps leaking ram
[QUOTE=Genericenemy;50802228]I think its inevitable we move to 8k, even TV will (The cameras already exist and have been used to film the Olympics in 2012) but probably not for a good while, 4K is bad enough as it is and now we will have to look at rendering 4X that and I don't see it happening currently. I'm personally yet to be convinced to move to even 4K would benefit my video games as much as 1080p did and would rather have a 1080p120 screen which would also give me the benefit of supporting 30FPS and 24FPS without dodgy framerate pulldown conversion.[/QUOTE]
To be honest, it's a bit interesting just to think about future tech. i remember when i was active on trubritars forums back in 2007-8 they were talking about how we'd probably see 12GB Gpus even up to 32GB in 10 years. to be honest trubritar was pretty spot on with his bet that 12GB cpus with 2ghz clocks would be the norm. it's a shame he's out of the loop now.
8k seems a bit overkill to me though, i never see any games at 4k even have artifacting on my girlfriend's parents 120hz 4k samsung that's 70". 8k would be ultimate for that, but i think VR will be the next leap in display tech, the problem is we need more powerful gpus and cheaper headsets for people. once the cost of a new headset like the vive drops to 200$, that's when we'll see an insane adoption rate, and people will have to keep up.
[editline]31st July 2016[/editline]
[QUOTE=TrafficMan;50803434]Good luck, phone repair was stressful but not as difficult as I thought it would be[/QUOTE]
i fixed my girlfriends galaxy s4 a few years ago and hilariously, i ripped out a wifi antenna and afterwards the internet was actually faster.
[QUOTE=Van-man;50803087]HP have a burning hatred for third party chargers.
I had to gut a (dead) official charger for it's ID chip so it'll work with my universal car charger for my old Pavillion DV7.
The bonus was that I could also use the official chargers cable which was also of better quality than the universal one's adapter.[/QUOTE]
Odd, I helped a friend order a dv6 charger replacement off Amazon since his died. It wasn't oem and it worked fine. Maybe it had a drm chip or something.
Sooooo... All the lights in the house just went out for a solid 15 seconds and my modem, which isn't on a UPS didn't disconnect. Just kept playing my game as normal. My router didn't even trip, and it's temperamental about brown outs.
It's not like the lights in my room went out. Every clock in the house got reset too.
:huh:
I was super proud of myself for figuring out which ethernet wire was which when we forgot to label them ( we had limited keystones, and 0 testing equipment)
They had numbers indicating how many feet you were on in terms of the 1000 ft roll.
I just said "hey John, how long was that chase?"
"100 feet, how's that gonna help?"
I see 444 on one of the wires and mark it, and follow it down and look for 544 on the other end. Marked it.
We used our last two keystones on that one wire that we needed to work the most out of any, and it was correct
[editline]31st July 2016[/editline]
so who else is following the thread where that one guy is screaming at everyone about how eugenics is the way to go
[QUOTE=Snickerdoodle;50804445]I was super proud of myself for figuring out which ethernet wire was which when we forgot to label them ( we had limited keystones, and 0 testing equipment)
They had numbers indicating how many feet you were on in terms of the 1000 ft roll.
I just said "hey John, how long was that chase?"
"100 feet, how's that gonna help?"
I see 444 on one of the wires and mark it, and follow it down and look for 544 on the other end. Marked it.
We used our last two keystones on that one wire that we needed to work the most out of any, and it was correct
[editline]31st July 2016[/editline]
so who else is following the thread where that one guy is screaming at everyone about how eugenics is the way to go[/QUOTE]
that reminds me that one time i was in anthropology class and my professor started talking about how good bacterial DNA can change and morph into different strains that are more or less deadly and how antibiotics aren't useful anymore. i brought up eugenics without mentioning it, saying that because antibiotics as well as other medicine are now available it makes the population weaker, mostly because of it's effect on saving people's lives whom would have died anyways, and a lot of the class actually were agreeing with me until i got to the point where i asked if it would be better if we let nature sort out the weak and dying. I've never seen any of my professors dodge an answer and end class that fast.
of course i was playing devils avocado so i don't really believe in eugenics.
also side note, a lot of the people who agreed with me took the same bus i did and wanted to abolish minimum wage because it makes people "Lazy"
Sorry, you need to Log In to post a reply to this thread.