Search Results

Search found 1325 results on 53 pages for 'factor'.

Page 22/53 | < Previous Page | 18 19 20 21 22 23 24 25 26 27 28 29  | Next Page >

  • What is an achievable way of setting content budgets (e.g. polygon count) for level content in a 3D title?

    - by MrCranky
    In answering this question for swquinn, the answer raised a more pertinent question that I'd like to hear answers to. I'll post our own strategy (promise I won't accept it as the answer), but I'd like to hear others. Specifically: how do you go about setting a sensible budget for your content team. Usually one of the very first questions asked in a development is: what's our polygon budget? Of course, these days it's rare that vertex/poly count alone is the limiting factor, instead shader complexity, fill-rate, lighting complexity, all come into play. What the content team want are some hard numbers / limits to work to such that they have a reasonable expectation that their content, once it actually gets into the engine, will not be too heavy. Given that 'it depends' isn't a particularly useful answer, I'd like to hear a strategy that allows me to give them workable limits without being a) misleading, or b) wrong.

    Read the article

  • Ask The Readers: How Do You Find Your Next Game?

    - by Jason Fitzpatrick
    Once upon a time the only place for new video game information was down at the arcade. These days there’s a news source and niche for everyone and every kind of game; where and how do you find your next video game conquest? Word of mouth? App recommendations? Critical reviews? This week we’re interested in lining ourselves up for a little summer fun: tell us all about your tips, tricks, and techniques for finding the real gems in the pile of games that comes out every year. Sound off in the comments and then check back in on Friday for the What You Said roundup. HTG Explains: What Is Two-Factor Authentication and Should I Be Using It? HTG Explains: What Is Windows RT and What Does It Mean To Me? HTG Explains: How Windows 8′s Secure Boot Feature Works & What It Means for Linux

    Read the article

  • .xprofile isn't enough

    - by BrianXP7
    I've been trying to find a solution on how to change the screen resolution on the login screen and other users. .xprofile will only affect my account. I've been searching for a few months but I got nothing. Please help. It would be easier if the "Default Resolution" was still there in the monitors settings. Plus, I'm afraid of editing the xorg.conf. Last time was ugly... My specs: HP dc5000 Small Form Factor Ubuntu 11.10 Oneiric Ocelot i386 Intel Pentium 4 2.40 GHz Intel 82865g x86/MMX/SSE2 Integrated Graphics (Standard Experience, OpenGL 1.3 and Runs Unity 3D :P) SoundMax Integrated Audio Card Broadcomm Integrated Ethernet 994.1 MiB RAM 38.3 GB HDD Acer X203H (Maximum Resolution 1600x900)

    Read the article

  • Why would i need extra IP adresses from my web host?

    - by user4524
    I am moving to a new and cheaper host. My old one raised the prices suddenly by a factor 10. Now what I did not like about the old one, was that each time I set up a new website, I had to set up a new account, with the result that I have to pay for a lot of webspace and a lot of bandwith i don't use. Now I am moving over to a new host, who cheaply offers virtual servers. Now I am a n00b when it comes to server tech. I have two questions about this: 1. Could I install all my webistes on this virtual server, provided it is big enough? 2. They offer the possibility of getting more than one IP address, but it costs more. What would be the advantage of this? Would I be able to suffice with one IP address?

    Read the article

  • Exploring TCP throughput with DTrace (2)

    - by user12820842
    Last time, I described how we can use the overlap in distributions of unacknowledged byte counts and send window to determine whether the peer's receive window may be too small, limiting throughput. Let's combine that comparison with a comparison of congestion window and slow start threshold, all on a per-port/per-client basis. This will help us Identify whether the congestion window or the receive window are limiting factors on throughput by comparing the distributions of congestion window and send window values to the distribution of outstanding (unacked) bytes. This will allow us to get a visual sense for how often we are thwarted in our attempts to fill the pipe due to congestion control versus the peer not being able to receive any more data. Identify whether slow start or congestion avoidance predominate by comparing the overlap in the congestion window and slow start distributions. If the slow start threshold distribution overlaps with the congestion window, we know that we have switched between slow start and congestion avoidance, possibly multiple times. Identify whether the peer's receive window is too small by comparing the distribution of outstanding unacked bytes with the send window distribution (i.e. the peer's receive window). I discussed this here. # dtrace -s tcp_window.d dtrace: script 'tcp_window.d' matched 10 probes ^C cwnd 80 10.175.96.92 value ------------- Distribution ------------- count 1024 | 0 2048 | 4 4096 | 6 8192 | 18 16384 | 36 32768 |@ 79 65536 |@ 155 131072 |@ 199 262144 |@@@ 400 524288 |@@@@@@ 798 1048576 |@@@@@@@@@@@@@@@@@@@@@@@@@@@@ 3848 2097152 | 0 ssthresh 80 10.175.96.92 value ------------- Distribution ------------- count 268435456 | 0 536870912 |@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@ 5543 1073741824 | 0 unacked 80 10.175.96.92 value ------------- Distribution ------------- count -1 | 0 0 | 1 1 | 0 2 | 0 4 | 0 8 | 0 16 | 0 32 | 0 64 | 0 128 | 0 256 | 3 512 | 0 1024 | 0 2048 | 4 4096 | 9 8192 | 21 16384 | 36 32768 |@ 78 65536 |@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@ 5391 131072 | 0 swnd 80 10.175.96.92 value ------------- Distribution ------------- count 32768 | 0 65536 |@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@ 5543 131072 | 0 Here we are observing a large file transfer via http on the webserver. Comparing these distributions, we can observe: That slow start congestion control is in operation. The distribution of congestion window values lies below the range of slow start threshold values (which are in the 536870912+ range), so the connection is in slow start mode. Both the unacked byte count and the send window values peak in the 65536-131071 range, but the send window value distribution is narrower. This tells us that the peer TCP's receive window is not closing. The congestion window distribution peaks in the 1048576 - 2097152 range while the receive window distribution is confined to the 65536-131071 range. Since the cwnd distribution ranges as low as 2048-4095, we can see that for some of the time we have been observing the connection, congestion control has been a limiting factor on transfer, but for the majority of the time the receive window of the peer would more likely have been the limiting factor. However, we know the window has never closed as the distribution of swnd values stays within the 65536-131071 range. So all in all we have a connection that has been mildly constrained by congestion control, but for the bulk of the time we have been observing it neither congestion or peer receive window have limited throughput. Here's the script: #!/usr/sbin/dtrace -s tcp:::send / (args[4]-tcp_flags & (TH_SYN|TH_RST|TH_FIN)) == 0 / { @cwnd["cwnd", args[4]-tcp_sport, args[2]-ip_daddr] = quantize(args[3]-tcps_cwnd); @ssthresh["ssthresh", args[4]-tcp_sport, args[2]-ip_daddr] = quantize(args[3]-tcps_cwnd_ssthresh); @unacked["unacked", args[4]-tcp_sport, args[2]-ip_daddr] = quantize(args[3]-tcps_snxt - args[3]-tcps_suna); @swnd["swnd", args[4]-tcp_sport, args[2]-ip_daddr] = quantize((args[4]-tcp_window)*(1 tcps_snd_ws)); } One surprise here is that slow start is still in operation - one would assume that for a large file transfer, acknowledgements would push the congestion window up past the slow start threshold over time. The slow start threshold is in fact still close to it's initial (very high) value, so that would suggest we have not experienced any congestion (the slow start threshold is adjusted when congestion occurs). Also, the above measurements were taken early in the connection lifetime, so the congestion window did not get a changes to get bumped up to the level of the slow start threshold. A good strategy when examining these sorts of measurements for a given service (such as a webserver) would be start by examining the distributions above aggregated by port number only to get an overall feel for service performance, i.e. is congestion control or peer receive window size an issue, or are we unconstrained to fill the pipe? From there, the overlap of distributions will tell us whether to drill down into specific clients. For example if the send window distribution has multiple peaks, we may want to examine if particular clients show issues with their receive window.

    Read the article

  • Writing a spell checker similar to "did you mean"

    - by user888734
    I'm hoping to write a spellchecker for search queries in a web application - not unlike Google's "Did you mean?" The algorithm will be loosely based on this: http://catalog.ldc.upenn.edu/LDC2006T13 In short, it generates correction candidates and scores them on how often they appear (along with adjacent words in the search query) in an enormous dataset of known n-grams - Google Web 1T - which contains well over 1 billion 5-grams. I'm not using the Web 1T dataset, but building my n-gram sets from my own documents - about 200k docs, and I'm estimating tens or hundreds of millions of n-grams will be generated. This kind of process is pushing the limits of my understanding of basic computing performance - can I simply load my n-grams into memory in a hashtable or dictionary when the app starts? Is the only limiting factor the amount of memory on the machine? Or am I barking up the wrong tree? Perhaps putting all my n-grams in a graph database with some sort of tree query optimisation? Could that ever be fast enough?

    Read the article

  • As a programmer, are you required to do timesheets?

    - by vcsjones
    Timesheets are something that I've never been fond of, but non-the-less something that is a requirement within my company. They don't bother me so much, but they seem to really grind some other people's gears. I suppose I have a few questions, and feedback would be great. Are you required to do timesheets, assuming you aren't a contractor? (That is understandable to me). What is the granularity of timesheets that you would be comfortable with or that you use? (ex: all entries must be under two hours). Would timesheets ever factor into your reasons for not accepting a job or leaving a current one? How has management within your organization justified timesheets if you aren't billing to a client?

    Read the article

  • Enjoy a Dazzling Desktop with the Brazil Theme for Windows 7

    - by Asian Angel
    Do you love a combination of nature and night-time city photography for your desktop? Then you will definitely want to download a copy of the Brazil Theme for Windows 7. The theme comes with six images featuring the colorful and unique beauty of Brazil. Download the Brazil Theme for Windows 7 [Windows 7 Personalization Gallery] How to Make Your Laptop Choose a Wired Connection Instead of Wireless HTG Explains: What Is Two-Factor Authentication and Should I Be Using It? HTG Explains: What Is Windows RT and What Does It Mean To Me?

    Read the article

  • SQLBits 9 goes to Liverpool

    SQLBits 9 has now been announced as 29th September to 1st October 2011 at the Adelphi Hotel in Liverpool. This will follow the now familiar three day format, a training day, and two full days of concurrent sessions. Saturday 1st October will of course be the free community day. Despite growing the event quite dramatically over the past four years, this is something we are all very proud to have maintained and is a key factor when planning the events. Plenty of more info to come, but in the meantime session submission is now open so why not submit an abstract?

    Read the article

  • Why You Should Choose A Web Development Company?

    Web development outsourcing has become the trend from last few years. In the past people were skeptical in sending work to other countries but in the present time this trend has become a boon. Web service outsourcing has become the important cost saving factor for the small sized businesses. But it also includes some risks so a proper choice of Web Service Company will maintain the balance between reducing the cost and minimizing the risks involved. India, China, Russia and Philippines have gained so much popularity for the web development services and SEO services.

    Read the article

  • Targeting a vehicle with complex movement?

    - by e100
    Targeting a vehicle with known constant velocity is simple, and collision is guaranteed. Imprecise AI can be modeled by adding a small error factor. But how would one go about targeting a vehicle whose movements are more complex? Perhaps it's evading the AI or another game object. I've been thinking about how I'd do it myself in a FPS (in which bullets have finite speed) and think there might need to be at least couple of targeting modes based on the target's movement in the previous second or so: If it's near linear (peak acceleration in a certain range) target with the linear model If it's highly irregular (perhaps size of bounding box of recent positions could be used?) , target at an average For now I can assume 2d space, AI is stationary and projectile moves linearly.

    Read the article

  • HTML5 / Native / C# & Mono [closed]

    - by iainjames88
    My apologies for the subjective nature of this question but I'm unsure as to which path to follow. I would like to do a bit of indie game development for the iPhone (nothing serious, just something I've wanted to pursue). At my university we aren't taught Java or Objective-C but C#/.NET and HTML5/JavaScript. Is it worth taking what I already know and try to accomplish my goal using, for example, C# and Mono or should I invest the time and learn Objective-C? I don't have a problem learning something new alongside my course (I love learning new stuff) and time isn't a factor. I'm slightly in favor of learning Objective-C for as it would be another string to my bow in the workplace, but it would be nice to stay with C# because it is what I'm used to.

    Read the article

  • Google Voice API for .NET

    - by lennykean
    There is no official API for Google Voice. However, many of the operations are done via simple rest style http calls, so building one shouldn’t be too difficult. Digging deeper and rolling your own is always a great way to exercise your problem solving skills and learn new things. So I did just that. As of the time of writing this, I’m only implementing basic SMS send since that’s all I needed. I’m hoping to implement more functionality in the future. In my next post, I’ll be using the SMS functionality to implement 2 factor authentication with ASP.Net MVC.

    Read the article

  • Does a large (hidden) submenu count towards site content in tems of determining page similarities?

    - by Name
    Basically, I have this site that recently lost a lot of traffic after I optimized the html, the exact reasons to which are uncertain. The graph of impressions (times a page appears on search listings) is continuously going down like an e^-x function. Because the content, previously occupying five pages of tables, now fits within a few paragraph tags, the menu now occupies about 80% of the live html code and I am starting to have doubts wherether this affects the "similar pages" factor that Google punishes. Questions: As far as I know, Google ignores invisible material and the submenus are only visible when hovered over. Has anything at all changed in this area? If I ajax in the submenus, leaving only the main eight menu items to load, will I be punished for "hiding" information? Is the idea worth testing or is it frankly retarded?

    Read the article

  • Producing JSON Documents from SQL Server queries via TSQL

    Although SQL Server supports XML well, XML's little cousin JSON gets no love. This is frustrating now that JSON is in so much demand. Maybe, Phil Factor suggests, it is possible to leverage all that XML, and XPath, goodness in SQL Server to produce JSON in a versatile way from SQL Queries? Yes, it so happens that there are plenty of alternatives. FREE eBook – "45 Database Performance Tips for Developers"Improve your database performance with 45 tips from SQL Server MVPs and industry experts. Get the eBook here.

    Read the article

  • How would I go about updating my electronic circuit simulator's 'electricity'?

    - by liqwidice
    I have made an application which allows the user to place down wires, power sources, and inverters on a virtual circuit board. All connections between tiles are automatic, as shown here: As you can see in the last image, the updating of power throughout this grid is not yet functioning. I think I understand how to do this updating conceptually, but getting it to work in my program is appearing to be much more difficult than I first imagined. My source code can be found here. If you have any tips as to how to I might approach this monstrous task, please let me know. EDIT The goal here is to simply get a working application. Getting tiles to pass on power to their neighbors is quite easy, but the tricky part is getting wires to "unpower" after the removal of a power source. The size of the grid is just 18x18, so efficiency really isn't a factor, for those wondering.

    Read the article

  • ????????

    - by ???02
    ????????Oracle Database VaultOracle Database Vault???????2???????????????????(DBA)???????????·?????????????????????????????????????????????????????????????????????????????????????????????????????????????????? ???????????????????????????????(??????)??????????????????????? Oracle Database Vault????????????DBA????????????????????DBA????????????????????????????? ??????????????????????????????????????Oracle Database Vault??Oracle-EBS?PeopleSoft?Siebel?SAP?????????????????????????????·?????????????????1. Separation of Duties (????)DBA(?????????)????????????????????????????????????????????????????·????? ????·???????????????????????????DBA????????????????????????????????????? ??????????????????Database Vault??DBA?????????????????????????????????????????????????????????????? ?????????????????????????????????????????2. ???(????)???????·?????????????·???????????????????????????????????????????????????????????????????????????????????????????????????????? SELECT/DML/EXECUTE ???????????????????????????????????????????????????????????????????????? DDL ???????????????????????????3. Multi-Factor Access Control Policy Enforcement (???????·??????)Oracle Database Vault???????????????????????????????????????????????????????·???????????????????????????????????????????????????DDL?DML???????????????????????????4. ???????·?????? ???????????????? (ex ???IP??????)??????? ??????????????? "??" or "???" ????????????????1???????????????????? ???·??? ???????????????? ???????????5. ????·???SQL???????????·???????????????·????true ???????????SQL??????????????????? ?????? Oracle Direct

    Read the article

  • Using logarithms to normalize a vector to avoid overflow

    - by muscicapa
    http://stackoverflow.com/questions/2293762/problem-with-arithmetic-using-logarithms-to-avoid-numerical-underflow-take-2 Having seen the above and having seen softmax normalization I was trying to normalize a vector while avoiding overflow - that is (x1 x2 x3 x4 ... xn) the normalized form for me has the sum of squares as 1.0 So what I thought of doing is s=(2*log(x1)+2*log(x2)+...+2*log(xn))/2 so the two factor can be taken off and finally the normalized vector is exp(log(x1)-s), , ..., exp(log(xn)-s) but I am evidently doing something wrong here, what?

    Read the article

  • Show frequencies along with barplot in ggplot2

    - by aL3xa
    I'm trying to display frequencies within barplot ... well, I want them somewhere in the graph: under the bars, within bars, above bars or in the legend area. And I recall (I may be wrong) that it can be done in ggplot2. This is probably an easy one... at least it seems easy. Here's the code: p <- ggplot(mtcars) p + aes(factor(cyl)) + geom_bar() Is there any chance that I can get frequencies embedded in the graph?

    Read the article

  • Get size of UIView after applying CGAffineTransform

    - by Ican Zilb
    I was surprised not to find an answer to this question, maybe is something very simple I somehow overlook : How to get the real size of an UIView after I apply a CGAffineTransform to it? eg. my UIView has size 300 x 200, I apply a scaling transform let's say factor 2 both horizontal and vertical, so the UIView now takes 600 x 400 on the screen, but it's bounds and it's layer's bounds are still returning a size of 300 x 200 ... where do I find the real size of the UIView ?

    Read the article

  • Drupal Query builder

    - by Rimian
    I quite often use Drupal's Views Module to build SQL that I paste into my code. It understands the Drupal database schema quite well. Is there a module that would give me this functionality or can I factor this out of Views?

    Read the article

  • Performance Optimization for Matrix Rotation

    - by Summer_More_More_Tea
    Hello everyone: I'm now trapped by a performance optimization lab in the book "Computer System from a Programmer's Perspective" described as following: In a N*N matrix M, where N is multiple of 32, the rotate operation can be represented as: Transpose: interchange elements M(i,j) and M(j,i) Exchange rows: Row i is exchanged with row N-1-i A example for matrix rotation(N is 3 instead of 32 for simplicity): ------- ------- |1|2|3| |3|6|9| ------- ------- |4|5|6| after rotate is |2|5|8| ------- ------- |7|8|9| |1|4|7| ------- ------- A naive implementation is: #define RIDX(i,j,n) ((i)*(n)+(j)) void naive_rotate(int dim, pixel *src, pixel *dst) { int i, j; for (i = 0; i < dim; i++) for (j = 0; j < dim; j++) dst[RIDX(dim-1-j, i, dim)] = src[RIDX(i, j, dim)]; } I come up with an idea by inner-loop-unroll. The result is: Code Version Speed Up original 1x unrolled by 2 1.33x unrolled by 4 1.33x unrolled by 8 1.55x unrolled by 16 1.67x unrolled by 32 1.61x I also get a code snippet from pastebin.com that seems can solve this problem: void rotate(int dim, pixel *src, pixel *dst) { int stride = 32; int count = dim >> 5; src += dim - 1; int a1 = count; do { int a2 = dim; do { int a3 = stride; do { *dst++ = *src; src += dim; } while(--a3); src -= dim * stride + 1; dst += dim - stride; } while(--a2); src += dim * (stride + 1); dst -= dim * dim - stride; } while(--a1); } After carefully read the code, I think main idea of this solution is treat 32 rows as a data zone, and perform the rotating operation respectively. Speed up of this version is 1.85x, overwhelming all the loop-unroll version. Here are the questions: In the inner-loop-unroll version, why does increment slow down if the unrolling factor increase, especially change the unrolling factor from 8 to 16, which does not effect the same when switch from 4 to 8? Does the result have some relationship with depth of the CPU pipeline? If the answer is yes, could the degrade of increment reflect pipeline length? What is the probable reason for the optimization of data-zone version? It seems that there is no too much essential difference from the original naive version. EDIT: My test environment is Intel Centrino Duo processor and the verion of gcc is 4.4 Any advice will be highly appreciated! Kind regards!

    Read the article

  • What is best approach for connection pooling?

    - by Bhushan
    I am implementing connection pooling in project. Performance wise which is better approach to do it? Hibernate (using C3PO or DBCP) Configuring JDBC data-source in Application server. Application server Portability is not an important factor for me. Please suggest the approach.

    Read the article

  • Scalability comparison between different DBMSs

    - by Björn Lindfors
    By what factor does the performance (read queries/sec) increase when a machine is added to a cluster of machines running either: a Bigtable-like database MySQL? Google's research paper on Bigtable suggests that "near-linear" scaling is achieved can be achieved with Bigtable. This page here featuring MySQL's marketing jargon suggests that MySQL is capable of scaling linearly. Where is the truth?

    Read the article

  • work benefits package [closed]

    - by Francisco Garcia
    For those of you who are into programming not just for the money. I would like to know which benefits you would like to have (or already have). OK, maybe taking away the money factor will limit this question too much. I am surprised to see that most companies have a fixed set for their benefits package. Were you able to negotiate something new or just your salary? What things have you seen out there and/or value most?

    Read the article

< Previous Page | 18 19 20 21 22 23 24 25 26 27 28 29  | Next Page >