## unity – “The mesh of Body.003 has 1 sub meshes but the renderer is using 4 materials. Your mesh should use the same amount of sub meshes as materials.”

I’m trying to import a Mesh (VRChat avatar) from Blender into Unity.

This mesh which is failing to import properly into Unity is separated in Blender as Head, Hair, Body. It looks great in Blender but imports into Unity as a floating head– as in, the Body is invisible so can only see Hair and Head.

The error given in the Unity log is the title of this post.

I don’t know what to actually do to fix it. When I Google, I can’t really comprehend any of the answers. Thanks for any suggestions.

## select – Average amount of orders per hour in the last X days per merchant

I have the following table named Orders:

``````MerchantID (INT)
OrderID (VARCHAR2)
OrderAmount (INT)
CreatedDate (DATE/TIME)
``````

I need to create a query that returns the average amount of orders per hour in the last X days per merchant.

While I do have some idea on how to get an average for the amount of orders, I can’t seem to combine all the requests for this particular item.

## matrices – How should I evaluate time complexity for matrix if I have a fixed (constant) amount of rows and columns?

Suppose, that I have a four-by-four matrix and I want to print each element of it.

``````matrix = ((1, 2, 3, 4), (5, 6, 7, 8), (9, 10, 11, 12), (13, 14, 15, 16))
for row in matrix
for elem in row
print(elem)
``````

So, I have a questions:

a) Should I consider, that such iterating requires O(k + n) in terms of big O notation, where k is a number of rows and n is a number of columns? I mean, $$sum_{i=1}^k1$$ + $$sum_{j=1}^n1$$ = O(k + n), we sum number of iterations that are required for rows and number of ones for columns. If I should not, then what is wrong with my estimates or how should I calculate big O for a matrix?

b) Can not we say, that such algorithm requires constant amount of time, because we have a well-defined input – four-by-for matrix, can we? I would like to specify what I mean: if we have constant input, does it mean, that our algo requires constant amount of time to compute in terms of big O?

## How to move a large amount of files off a SharePoint 2003 Portal Server to SharePoint Online?

One obvious solution would be to upgrade up to a point where either SPMT or Migration Manager could take over the migration process.
I believe that the lower boundary for the above tools is that the Sharepoint Server is at least 10. So that would mean that you would have to upgrade from 2003 to moss 2007 and from there to sharepoint 2010.

Another solution, would be to check out the third party tools and their compatibility with the SPS 2003(Honestly, i do not think that there is still any tool that could handle such a task).

A third option is something that i used in order to migrate a portion of files from our own on-premises moss 2010 to out online tennant, just to see if it is applicable.
I am not sure if you can use this solution, but i will type it down so that you can try it out.

Moss 2010 gave you the option to open a Document Library as a network mapped drive.
So what i did was :

1. open a network mapped drive on the fileserver, in which case a letter is assigned to the drive.
2. write a simple script with the robocopy command just to utilise the parallel functionality of cloning folders that it has in order to get the files that i needed in the FileServer.
3. Use SPMT that pointed to that file location and upload all of the data.

If in any case you wish to proceed with testing out this solution, i would like to stress out that while copying files over the network with the parallel option, in robocopy, enabled did create a huge load to the sharepoint server. I’m just stating that because, i, also, attempted to upload files to an network mapped drive that pointed to a SharePoint Online Document Library. The result of the above was that the connection was cut off after a couple of minutes đź™‚

One of the drawbacks was that permissions and file modification dates were not preserved, but i believe that since you are trying to migrate something that is as old as SPS 2003, this will be the least of your concerns.

## module – Magento 2: Custom Order Date and Shipping Amount in Programmatically created Order

module – Magento 2: Custom Order Date and Shipping Amount in Programmatically created Order – Magento Stack Exchange

## means – Probability of observing a certain amount of arrivals in an interval

I’m told the mean rate of arrivals in a facility is 10 per minute.

I’m asked to find the probability that more than 2 arrivals occur within 10 seconds.

I know that $$P(>2$$ arrivals$$) = 1-P(<=2$$ arrivals$$)$$ but since the interval is 10 seconds and not 60 I’m a bit confused.

If they asked what the probability of more than 2 arrivals in 60 seconds was I could just use the Poisson table.

In order to answer this question do I just convert my mean rate? So instead of 10 per minute it could be $$10/6$$ arrivals per 10 seconds and then use the poisson table?

I’m just a bit confused on what to do when the mean rate is given as ‘per minute’ and the question is only looking at a 10 second interval.

I hope I’ve explained it well enough, any help would be appreciated.

## java – Architectural design for sending large amount of analytics data from production servers to s3 without impacting request performance

Lets say we have a server getting upto 1000 requests per second, serving them at p99 of 20ms (strong business case for not increasing this latency). The server gc parameters have been carefully tuned for this performance and current latency is already bottlenecked by gc. We want to log structured data related to requests and responses, ideally 100% of it without dropping anything, to S3 in for example gzipped jsonlines format (analytics will be done on this data, each file should be ideally 100MB-500MB in size). Analytics does not have to be realtime. A few hours of delay, for example, is fine. Also the IOUtilization already approaches 100% so writing this data to disk at any time is likely not an option. All code is in Java.

Solution 1:
Use the threads getting and serving requests as producers and have them enqueue each request/response into blocking buffer(s) with error/edge case handling of buffer being full, exception, etc. This way the producer threads dont get blocked no matter what. Then have a consumer threadpool consume from these buffer(s) in a batched way, compress and send to s3. The upside is that it is a simple(ish) solution. The main downside is that all this is done in the same jvm and might increase allocation rate and degrade performance for main requests? I suspect the main source of new object creation might be during serialization to string (is this true?). Putting objects into a fixed queue size or draining to (using drainTo method on BlockingQueue) to an existing collection should not allocate anything new I think.

Solution 2:
Setup a separate service running on the same host (so separate jvm with its own tuned gc if necessary) that exposes endpoints like locaholhost:8080/request for example. Producers send data to these endpoints and all consumer logic lies in this service (mostly same as before). Downside is that this might be more complex. Also sending data, even to localhost, might block the producer thread (whose main job is to serve requests) and decrease throughput per host?

For Solution 1 or 2 are there any Java compatible libraries (producer/consumer libraries or high performance TCP based messaging libraries) that might be appropriate to use instead of rolling my own?

I know these questions can be answered by benchmarking and making a poc, but looking for some direction in case someone has suggestions or maybe a third way I haven’t though of.

## magento2 – How to exclude certain SKU to be calculated in Minimum Order Amount

I’m currently using the native M2 Minimum Order Amount module.

I want to add to my store some products, but I don’t want them to be consider during minimum order amount calculation.

Minimum_Order_Amount = 30

Example
If cart contains some SKU’s
Then Minimum_Order_Amount += itemRowSubtotal

I’m wondering how to approach this problem?
Also I would like to return dynamic error equivalent to change.

## cognitive load – What’s the rationale behind Paypal’s amount input?

This is something I’ve always wanted to ask because it’s really frustrating for me.

If you use Paypal, you will see that when you try to send money, the Amount `input` requires to enter the cents or decimals. Look at the picture below

In this real case, I had to send \$160. On most apps, you enter \$160. On Paypal you have to enter 16000 or you send \$1.60. I’ve had this error once or twice, and now I’m very careful to double and triple check because I sometimes enter an extra 0.

I’ve never seen this before (although I’ve started to see this AFTER Paypal started to do it ). Based on the principles of intentionality in UX, if I wanted to add decimals, I would use floating point and then add decimals. So I’m wondering why they would do something so confusing and if there is some sort of rationale that I’m not aware of

Also, from a Universal UX theory perspective, this is a fail. Paypal knows my country and my language. So they should know that floating point has a whole different meaning for me (we use commas for decimals). Therefore, it’s an even bigger cognitive load. This is a small consideration in the grand scheme of things, but pretty striking when you consider such an important company.

Anyway, is there some kind of explanation from Paypal or a rationale I’m not getting (maybe because of cross-cultural barriers?)? Or is it just an anti-pattern?